From 650250ea8e44542a6f2cab43b4e6b8932e5906fa Mon Sep 17 00:00:00 2001 From: skunkworxdark Date: Sun, 5 Apr 2026 17:25:16 +0100 Subject: [PATCH 1/8] feat(model-manager): add comprehensive sorting capabilities for models dded the ability to sort models in the Model Manager by various attributes including Name, Base, Type, Format, Size, Date Added, and Date Modified. Supports both ascending and descending order. - Backend: Added `order_by` and `direction` query parameters to the ``/api/v1/models`/` listing endpoint. Implemented case-insensitive sorting in the SQLite model records service. - Frontend: Introduced `` UI, updated Redux slices to manage sort state, removed client-side entity adapter sorting to respect server-side ordering, and added i18n localization keys. - Tests: Added test coverage for SQL-based sorting on size and name. --- invokeai/app/api/routers/model_manager.py | 8 +- .../model_records/model_records_base.py | 13 ++- .../model_records/model_records_sql.py | 35 ++++++-- invokeai/frontend/web/public/locales/en.json | 9 ++ .../store/modelManagerV2Slice.ts | 16 ++++ .../subpanels/ModelManagerPanel/ModelList.tsx | 7 +- .../ModelManagerPanel/ModelListNavigation.tsx | 4 + .../ModelManagerPanel/ModelSortControl.tsx | 85 +++++++++++++++++++ .../web/src/services/api/endpoints/models.ts | 15 +++- .../model_records/test_model_records_sql.py | 67 +++++++++++++++ 10 files changed, 242 insertions(+), 17 deletions(-) create mode 100644 invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx diff --git a/invokeai/app/api/routers/model_manager.py b/invokeai/app/api/routers/model_manager.py index 65b059ecfce..f510cd8e382 100644 --- a/invokeai/app/api/routers/model_manager.py +++ b/invokeai/app/api/routers/model_manager.py @@ -26,8 +26,10 @@ from invokeai.app.services.model_records import ( InvalidModelException, ModelRecordChanges, + ModelRecordOrderBy, UnknownModelException, ) +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.services.orphaned_models import OrphanedModelInfo from invokeai.app.util.suppress_output import SuppressOutput from invokeai.backend.model_manager.configs.factory import AnyModelConfig, ModelConfigFactory @@ -130,6 +132,8 @@ async def list_model_records( model_format: Optional[ModelFormat] = Query( default=None, description="Exact match on the format of the model (e.g. 'diffusers')" ), + order_by: ModelRecordOrderBy = Query(default=ModelRecordOrderBy.Name, description="The field to order by"), + direction: SQLiteDirection = Query(default=SQLiteDirection.Ascending, description="The direction to order by"), ) -> ModelsList: """Get a list of models.""" record_store = ApiDependencies.invoker.services.model_manager.store @@ -138,12 +142,12 @@ async def list_model_records( for base_model in base_models: found_models.extend( record_store.search_by_attr( - base_model=base_model, model_type=model_type, model_name=model_name, model_format=model_format + base_model=base_model, model_type=model_type, model_name=model_name, model_format=model_format, order_by=order_by, direction=direction ) ) else: found_models.extend( - record_store.search_by_attr(model_type=model_type, model_name=model_name, model_format=model_format) + record_store.search_by_attr(model_type=model_type, model_name=model_name, model_format=model_format, order_by=order_by, direction=direction) ) for model in found_models: model = add_cover_image_to_model_config(model, ApiDependencies) diff --git a/invokeai/app/services/model_records/model_records_base.py b/invokeai/app/services/model_records/model_records_base.py index 96e12d3b0a3..694718715ba 100644 --- a/invokeai/app/services/model_records/model_records_base.py +++ b/invokeai/app/services/model_records/model_records_base.py @@ -11,6 +11,7 @@ from pydantic import BaseModel, Field from invokeai.app.services.shared.pagination import PaginatedResults +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.util.model_exclude_null import BaseModelExcludeNull from invokeai.backend.model_manager.configs.controlnet import ControlAdapterDefaultSettings from invokeai.backend.model_manager.configs.factory import AnyModelConfig @@ -55,6 +56,10 @@ class ModelRecordOrderBy(str, Enum): Base = "base" Name = "name" Format = "format" + Size = "size" + DateAdded = "created_at" + DateModified = "updated_at" + Path = "path" class ModelSummary(BaseModel): @@ -178,7 +183,11 @@ def get_model_by_hash(self, hash: str) -> AnyModelConfig: @abstractmethod def list_models( - self, page: int = 0, per_page: int = 10, order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default + self, + page: int = 0, + per_page: int = 10, + order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default, + direction: SQLiteDirection = SQLiteDirection.Ascending, ) -> PaginatedResults[ModelSummary]: """Return a paginated summary listing of each model in the database.""" pass @@ -215,6 +224,8 @@ def search_by_attr( base_model: Optional[BaseModelType] = None, model_type: Optional[ModelType] = None, model_format: Optional[ModelFormat] = None, + order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default, + direction: SQLiteDirection = SQLiteDirection.Ascending, ) -> List[AnyModelConfig]: """ Return models matching name, base and/or type. diff --git a/invokeai/app/services/model_records/model_records_sql.py b/invokeai/app/services/model_records/model_records_sql.py index edcbba2acdc..f104c3855e7 100644 --- a/invokeai/app/services/model_records/model_records_sql.py +++ b/invokeai/app/services/model_records/model_records_sql.py @@ -57,6 +57,7 @@ UnknownModelException, ) from invokeai.app.services.shared.pagination import PaginatedResults +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.services.shared.sqlite.sqlite_database import SqliteDatabase from invokeai.backend.model_manager.configs.factory import AnyModelConfig, ModelConfigFactory from invokeai.backend.model_manager.taxonomy import BaseModelType, ModelFormat, ModelType @@ -257,6 +258,7 @@ def search_by_attr( model_type: Optional[ModelType] = None, model_format: Optional[ModelFormat] = None, order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default, + direction: SQLiteDirection = SQLiteDirection.Ascending, ) -> List[AnyModelConfig]: """ Return models matching name, base and/or type. @@ -266,18 +268,24 @@ def search_by_attr( :param model_type: Filter by type of model (optional) :param model_format: Filter by model format (e.g. "diffusers") (optional) :param order_by: Result order + :param direction: Result direction If none of the optional filters are passed, will return all models in the database. """ with self._db.transaction() as cursor: assert isinstance(order_by, ModelRecordOrderBy) + order_dir = "DESC" if direction == SQLiteDirection.Descending else "ASC" ordering = { - ModelRecordOrderBy.Default: "type, base, name, format", + ModelRecordOrderBy.Default: f"type {order_dir}, base COLLATE NOCASE {order_dir}, name COLLATE NOCASE {order_dir}, format", ModelRecordOrderBy.Type: "type", - ModelRecordOrderBy.Base: "base", - ModelRecordOrderBy.Name: "name", + ModelRecordOrderBy.Base: "base COLLATE NOCASE", + ModelRecordOrderBy.Name: "name COLLATE NOCASE", ModelRecordOrderBy.Format: "format", + ModelRecordOrderBy.Size: "IFNULL(json_extract(config, '$.file_size'), 0)", + ModelRecordOrderBy.DateAdded: "created_at", + ModelRecordOrderBy.DateModified: "updated_at", + ModelRecordOrderBy.Path: "path", } where_clause: list[str] = [] @@ -301,7 +309,7 @@ def search_by_attr( SELECT config FROM models {where} - ORDER BY {ordering[order_by]} -- using ? to bind doesn't work here for some reason; + ORDER BY {ordering[order_by]} {order_dir} -- using ? to bind doesn't work here for some reason; """, tuple(bindings), ) @@ -357,17 +365,26 @@ def search_by_hash(self, hash: str) -> List[AnyModelConfig]: return results def list_models( - self, page: int = 0, per_page: int = 10, order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default + self, + page: int = 0, + per_page: int = 10, + order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default, + direction: SQLiteDirection = SQLiteDirection.Ascending, ) -> PaginatedResults[ModelSummary]: """Return a paginated summary listing of each model in the database.""" with self._db.transaction() as cursor: assert isinstance(order_by, ModelRecordOrderBy) + order_dir = "DESC" if direction == SQLiteDirection.Descending else "ASC" ordering = { - ModelRecordOrderBy.Default: "type, base, name, format", + ModelRecordOrderBy.Default: f"type {order_dir}, base COLLATE NOCASE {order_dir}, name COLLATE NOCASE {order_dir}, format", ModelRecordOrderBy.Type: "type", - ModelRecordOrderBy.Base: "base", - ModelRecordOrderBy.Name: "name", + ModelRecordOrderBy.Base: "base COLLATE NOCASE", + ModelRecordOrderBy.Name: "name COLLATE NOCASE", ModelRecordOrderBy.Format: "format", + ModelRecordOrderBy.Size: "IFNULL(json_extract(config, '$.file_size'), 0)", + ModelRecordOrderBy.DateAdded: "created_at", + ModelRecordOrderBy.DateModified: "updated_at", + ModelRecordOrderBy.Path: "path", } # Lock so that the database isn't updated while we're doing the two queries. @@ -385,7 +402,7 @@ def list_models( f"""--sql SELECT config FROM models - ORDER BY {ordering[order_by]} -- using ? to bind doesn't work here for some reason + ORDER BY {ordering[order_by]} {order_dir} -- using ? to bind doesn't work here for some reason LIMIT ? OFFSET ?; """, diff --git a/invokeai/frontend/web/public/locales/en.json b/invokeai/frontend/web/public/locales/en.json index 285dc0817e1..3fce8e2861d 100644 --- a/invokeai/frontend/web/public/locales/en.json +++ b/invokeai/frontend/web/public/locales/en.json @@ -1125,6 +1125,15 @@ "modelType": "Model Type", "modelUpdated": "Model Updated", "modelUpdateFailed": "Model Update Failed", + "sortByName": "Name", + "sortByBase": "Base", + "sortBySize": "Size", + "sortByDateAdded": "Date Added", + "sortByDateModified": "Date Modified", + "sortByPath": "Path", + "sortByType": "Type", + "sortByFormat": "Format", + "sortDefault": "Default", "name": "Name", "modelPickerFallbackNoModelsInstalled": "No models installed.", "modelPickerFallbackNoModelsInstalled2": "Visit the Model Manager to install models.", diff --git a/invokeai/frontend/web/src/features/modelManagerV2/store/modelManagerV2Slice.ts b/invokeai/frontend/web/src/features/modelManagerV2/store/modelManagerV2Slice.ts index 092998d0c31..b6aea4138bb 100644 --- a/invokeai/frontend/web/src/features/modelManagerV2/store/modelManagerV2Slice.ts +++ b/invokeai/frontend/web/src/features/modelManagerV2/store/modelManagerV2Slice.ts @@ -22,6 +22,10 @@ const zModelManagerState = z.object({ scanPath: z.string().optional(), shouldInstallInPlace: z.boolean(), selectedModelKeys: z.array(z.string()), + orderBy: z + .enum(['default', 'name', 'type', 'base', 'size', 'created_at', 'updated_at', 'path', 'format']) + .default('name'), + sortDirection: z.enum(['asc', 'desc']).default('asc'), }); type ModelManagerState = z.infer; @@ -35,6 +39,8 @@ const getInitialState = (): ModelManagerState => ({ scanPath: undefined, shouldInstallInPlace: true, selectedModelKeys: [], + orderBy: 'name', + sortDirection: 'asc', }); const slice = createSlice({ @@ -74,6 +80,12 @@ const slice = createSlice({ clearModelSelection: (state) => { state.selectedModelKeys = []; }, + setOrderBy: (state, action: PayloadAction) => { + state.orderBy = action.payload; + }, + setSortDirection: (state, action: PayloadAction) => { + state.sortDirection = action.payload; + }, }, }); @@ -87,6 +99,8 @@ export const { modelSelectionChanged, toggleModelSelection, clearModelSelection, + setOrderBy, + setSortDirection, } = slice.actions; export const modelManagerSliceConfig: SliceConfig = { @@ -116,3 +130,5 @@ export const selectSearchTerm = createModelManagerSelector((mm) => mm.searchTerm export const selectFilteredModelType = createModelManagerSelector((mm) => mm.filteredModelType); export const selectShouldInstallInPlace = createModelManagerSelector((mm) => mm.shouldInstallInPlace); export const selectSelectedModelKeys = createModelManagerSelector((mm) => mm.selectedModelKeys); +export const selectOrderBy = createModelManagerSelector((mm) => mm.orderBy); +export const selectSortDirection = createModelManagerSelector((mm) => mm.sortDirection); diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelList.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelList.tsx index efb5c1add22..165150dd7f4 100644 --- a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelList.tsx +++ b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelList.tsx @@ -8,8 +8,10 @@ import { clearModelSelection, type FilterableModelType, selectFilteredModelType, + selectOrderBy, selectSearchTerm, selectSelectedModelKeys, + selectSortDirection, setSelectedModelKey, } from 'features/modelManagerV2/store/modelManagerV2Slice'; import { memo, useCallback, useMemo, useState } from 'react'; @@ -39,6 +41,8 @@ const ModelList = () => { const dispatch = useAppDispatch(); const filteredModelType = useAppSelector(selectFilteredModelType); const searchTerm = useAppSelector(selectSearchTerm); + const orderBy = useAppSelector(selectOrderBy); + const direction = useAppSelector(selectSortDirection); const selectedModelKeys = useAppSelector(selectSelectedModelKeys); const { t } = useTranslation(); const toast = useToast(); @@ -47,7 +51,8 @@ const ModelList = () => { const [isDeleting, setIsDeleting] = useState(false); const [isReidentifying, setIsReidentifying] = useState(false); - const { data: allModelsData, isLoading: isLoadingAll } = useGetModelConfigsQuery(); + const queryArgs = useMemo(() => ({ order_by: orderBy, direction: direction.toUpperCase() }), [orderBy, direction]); + const { data: allModelsData, isLoading: isLoadingAll } = useGetModelConfigsQuery(queryArgs); const { data: missingModelsData, isLoading: isLoadingMissing } = useGetMissingModelsQuery(); const [bulkDeleteModels] = useBulkDeleteModelsMutation(); const [bulkReidentifyModels] = useBulkReidentifyModelsMutation(); diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx index 78bed8ab830..2f3dcb94efe 100644 --- a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx +++ b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx @@ -7,6 +7,7 @@ import { memo, useCallback } from 'react'; import { PiXBold } from 'react-icons/pi'; import { ModelListBulkActions } from './ModelListBulkActions'; +import { ModelSortControl } from './ModelSortControl'; import { ModelTypeFilter } from './ModelTypeFilter'; export const ModelListNavigation = memo(() => { @@ -49,6 +50,9 @@ export const ModelListNavigation = memo(() => { )} + + + diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx new file mode 100644 index 00000000000..1e84de30dbf --- /dev/null +++ b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx @@ -0,0 +1,85 @@ +import { Flex, IconButton, Select } from '@invoke-ai/ui-library'; +import { useAppDispatch, useAppSelector } from 'app/store/storeHooks'; +import { + selectOrderBy, + selectSortDirection, + setOrderBy, + setSortDirection, +} from 'features/modelManagerV2/store/modelManagerV2Slice'; +import type { ChangeEvent } from 'react'; +import { useCallback, useMemo } from 'react'; +import { useTranslation } from 'react-i18next'; +import { PiSortAscendingBold, PiSortDescendingBold } from 'react-icons/pi'; +import { z } from 'zod'; + +const zOrderBy = z.enum(['default', 'name', 'type', 'base', 'size', 'created_at', 'updated_at', 'path', 'format']); +type OrderBy = z.infer; +const isOrderBy = (v: unknown): v is OrderBy => zOrderBy.safeParse(v).success; + +const ORDER_BY_OPTIONS: OrderBy[] = [ + 'default', + 'name', + 'base', + 'size', + 'created_at', + 'updated_at', + 'path', + 'type', + 'format', +]; + +export const ModelSortControl = () => { + const { t } = useTranslation(); + const dispatch = useAppDispatch(); + + const orderBy = useAppSelector(selectOrderBy); + const direction = useAppSelector(selectSortDirection); + + const ORDER_BY_LABELS = useMemo( + () => ({ + default: t('modelManager.sortDefault'), + name: t('modelManager.sortByName'), + base: t('modelManager.sortByBase'), + size: t('modelManager.sortBySize'), + created_at: t('modelManager.sortByDateAdded'), + updated_at: t('modelManager.sortByDateModified'), + path: t('modelManager.sortByPath'), + type: t('modelManager.sortByType'), + format: t('modelManager.sortByFormat'), + }), + [t] + ); + + const onChangeOrderBy = useCallback( + (e: ChangeEvent) => { + if (!isOrderBy(e.target.value)) { + return; + } + dispatch(setOrderBy(e.target.value)); + }, + [dispatch] + ); + + const toggleDirection = useCallback(() => { + dispatch(setSortDirection(direction === 'asc' ? 'desc' : 'asc')); + }, [dispatch, direction]); + + return ( + + + : } + size="sm" + variant="ghost" + onClick={toggleDirection} + /> + + ); +}; diff --git a/invokeai/frontend/web/src/services/api/endpoints/models.ts b/invokeai/frontend/web/src/services/api/endpoints/models.ts index c3d0decd53c..f279d46d823 100644 --- a/invokeai/frontend/web/src/services/api/endpoints/models.ts +++ b/invokeai/frontend/web/src/services/api/endpoints/models.ts @@ -111,9 +111,13 @@ type DeleteOrphanedModelsResponse = { errors: Record; }; +type GetModelConfigsArg = { + order_by?: string; + direction?: string; +} | void; + const modelConfigsAdapter = createEntityAdapter({ selectId: (entity) => entity.key, - sortComparer: (a, b) => a.name.localeCompare(b.name), }); export const modelConfigsAdapterSelectors = modelConfigsAdapter.getSelectors(undefined, getSelectorsOptions); @@ -338,8 +342,11 @@ export const modelsApi = api.injectEndpoints({ }, invalidatesTags: ['ModelInstalls'], }), - getModelConfigs: build.query, void>({ - query: () => ({ url: buildModelsUrl() }), + getModelConfigs: build.query, GetModelConfigsArg>({ + query: (arg) => { + const queryStr = arg ? `?${queryString.stringify(arg)}` : ''; + return { url: buildModelsUrl(queryStr) }; + }, providesTags: (result) => { const tags: ApiTagDescription[] = [{ type: 'ModelConfig', id: LIST_TAG }]; if (result) { @@ -498,5 +505,5 @@ export const { useDeleteOrphanedModelsMutation, } = modelsApi; -export const selectModelConfigsQuery = modelsApi.endpoints.getModelConfigs.select(); +export const selectModelConfigsQuery = modelsApi.endpoints.getModelConfigs.select(undefined); export const selectMissingModelsQuery = modelsApi.endpoints.getMissingModels.select(); diff --git a/tests/app/services/model_records/test_model_records_sql.py b/tests/app/services/model_records/test_model_records_sql.py index 2b6c54d5b0f..1bf697ff4cd 100644 --- a/tests/app/services/model_records/test_model_records_sql.py +++ b/tests/app/services/model_records/test_model_records_sql.py @@ -14,7 +14,9 @@ ModelRecordServiceBase, ModelRecordServiceSQL, UnknownModelException, + ModelRecordOrderBy, ) +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.services.model_records.model_records_base import ModelRecordChanges from invokeai.backend.model_manager.configs.controlnet import ControlAdapterDefaultSettings from invokeai.backend.model_manager.configs.lora import LoRA_LyCORIS_SDXL_Config @@ -363,6 +365,71 @@ def test_filter_2(store: ModelRecordServiceBase): ) assert len(matches) == 1 +def test_search_by_attr_sorting(store: ModelRecordServiceSQL): + config1 = Main_Diffusers_SD1_Config( + path="/tmp/config1", + name="alpha", + base=BaseModelType.StableDiffusion1, + type=ModelType.Main, + hash="CONFIG1HASH", + file_size=1000, + source="test/source/", + source_type=ModelSourceType.Path, + variant=ModelVariantType.Normal, + prediction_type=SchedulerPredictionType.Epsilon, + repo_variant=ModelRepoVariant.Default, + ) + config2 = Main_Diffusers_SD2_Config( + path="/tmp/config2", + name="beta", + base=BaseModelType.StableDiffusion2, + type=ModelType.Main, + hash="CONFIG2HASH", + file_size=2000, + source="test/source/", + source_type=ModelSourceType.Path, + variant=ModelVariantType.Normal, + prediction_type=SchedulerPredictionType.Epsilon, + repo_variant=ModelRepoVariant.Default, + ) + config3 = VAE_Diffusers_SD1_Config( + path="/tmp/config3", + name="gamma", + base=BaseModelType.StableDiffusion1, + type=ModelType.VAE, + hash="CONFIG3HASH", + file_size=500, + source="test/source/", + source_type=ModelSourceType.Path, + repo_variant=ModelRepoVariant.Default, + ) + for c in config1, config2, config3: + store.add_model(c) + + # Test sorting by Name Ascending + matches = store.search_by_attr(order_by=ModelRecordOrderBy.Name, direction=SQLiteDirection.Ascending) + assert len(matches) == 3 + assert matches[0].name == "alpha" + assert matches[1].name == "beta" + assert matches[2].name == "gamma" + + # Test sorting by Name Descending + matches = store.search_by_attr(order_by=ModelRecordOrderBy.Name, direction=SQLiteDirection.Descending) + assert matches[0].name == "gamma" + assert matches[1].name == "beta" + assert matches[2].name == "alpha" + + # Test sorting by Size Ascending + matches = store.search_by_attr(order_by=ModelRecordOrderBy.Size, direction=SQLiteDirection.Ascending) + assert matches[0].name == "gamma" # 500 + assert matches[1].name == "alpha" # 1000 + assert matches[2].name == "beta" # 2000 + + # Test sorting by Size Descending + matches = store.search_by_attr(order_by=ModelRecordOrderBy.Size, direction=SQLiteDirection.Descending) + assert matches[0].name == "beta" # 2000 + assert matches[1].name == "alpha" # 1000 + assert matches[2].name == "gamma" # 500 def test_model_record_changes(): # This test guards against some unexpected behaviours from pydantic's union evaluation. See #6035 From cc231bcf8cfb004eeecc1d80c8fd59c6f07f5f1d Mon Sep 17 00:00:00 2001 From: skunkworxdark Date: Sun, 5 Apr 2026 17:25:16 +0100 Subject: [PATCH 2/8] feat(model-manager): add comprehensive sorting capabilities for models dded the ability to sort models in the Model Manager by various attributes including Name, Base, Type, Format, Size, Date Added, and Date Modified. Supports both ascending and descending order. - Backend: Added `order_by` and `direction` query parameters to the ``/api/v1/models`/` listing endpoint. Implemented case-insensitive sorting in the SQLite model records service. - Frontend: Introduced `` UI, updated Redux slices to manage sort state, removed client-side entity adapter sorting to respect server-side ordering, and added i18n localization keys. - Tests: Added test coverage for SQL-based sorting on size and name. --- invokeai/app/api/routers/model_manager.py | 8 +- .../model_records/model_records_base.py | 13 ++- .../model_records/model_records_sql.py | 35 ++++++-- invokeai/frontend/web/public/locales/en.json | 9 ++ .../store/modelManagerV2Slice.ts | 16 ++++ .../subpanels/ModelManagerPanel/ModelList.tsx | 7 +- .../ModelManagerPanel/ModelListNavigation.tsx | 4 + .../ModelManagerPanel/ModelSortControl.tsx | 85 +++++++++++++++++++ .../web/src/services/api/endpoints/models.ts | 15 +++- .../model_records/test_model_records_sql.py | 67 +++++++++++++++ 10 files changed, 242 insertions(+), 17 deletions(-) create mode 100644 invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx diff --git a/invokeai/app/api/routers/model_manager.py b/invokeai/app/api/routers/model_manager.py index 65b059ecfce..f510cd8e382 100644 --- a/invokeai/app/api/routers/model_manager.py +++ b/invokeai/app/api/routers/model_manager.py @@ -26,8 +26,10 @@ from invokeai.app.services.model_records import ( InvalidModelException, ModelRecordChanges, + ModelRecordOrderBy, UnknownModelException, ) +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.services.orphaned_models import OrphanedModelInfo from invokeai.app.util.suppress_output import SuppressOutput from invokeai.backend.model_manager.configs.factory import AnyModelConfig, ModelConfigFactory @@ -130,6 +132,8 @@ async def list_model_records( model_format: Optional[ModelFormat] = Query( default=None, description="Exact match on the format of the model (e.g. 'diffusers')" ), + order_by: ModelRecordOrderBy = Query(default=ModelRecordOrderBy.Name, description="The field to order by"), + direction: SQLiteDirection = Query(default=SQLiteDirection.Ascending, description="The direction to order by"), ) -> ModelsList: """Get a list of models.""" record_store = ApiDependencies.invoker.services.model_manager.store @@ -138,12 +142,12 @@ async def list_model_records( for base_model in base_models: found_models.extend( record_store.search_by_attr( - base_model=base_model, model_type=model_type, model_name=model_name, model_format=model_format + base_model=base_model, model_type=model_type, model_name=model_name, model_format=model_format, order_by=order_by, direction=direction ) ) else: found_models.extend( - record_store.search_by_attr(model_type=model_type, model_name=model_name, model_format=model_format) + record_store.search_by_attr(model_type=model_type, model_name=model_name, model_format=model_format, order_by=order_by, direction=direction) ) for model in found_models: model = add_cover_image_to_model_config(model, ApiDependencies) diff --git a/invokeai/app/services/model_records/model_records_base.py b/invokeai/app/services/model_records/model_records_base.py index 96e12d3b0a3..694718715ba 100644 --- a/invokeai/app/services/model_records/model_records_base.py +++ b/invokeai/app/services/model_records/model_records_base.py @@ -11,6 +11,7 @@ from pydantic import BaseModel, Field from invokeai.app.services.shared.pagination import PaginatedResults +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.util.model_exclude_null import BaseModelExcludeNull from invokeai.backend.model_manager.configs.controlnet import ControlAdapterDefaultSettings from invokeai.backend.model_manager.configs.factory import AnyModelConfig @@ -55,6 +56,10 @@ class ModelRecordOrderBy(str, Enum): Base = "base" Name = "name" Format = "format" + Size = "size" + DateAdded = "created_at" + DateModified = "updated_at" + Path = "path" class ModelSummary(BaseModel): @@ -178,7 +183,11 @@ def get_model_by_hash(self, hash: str) -> AnyModelConfig: @abstractmethod def list_models( - self, page: int = 0, per_page: int = 10, order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default + self, + page: int = 0, + per_page: int = 10, + order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default, + direction: SQLiteDirection = SQLiteDirection.Ascending, ) -> PaginatedResults[ModelSummary]: """Return a paginated summary listing of each model in the database.""" pass @@ -215,6 +224,8 @@ def search_by_attr( base_model: Optional[BaseModelType] = None, model_type: Optional[ModelType] = None, model_format: Optional[ModelFormat] = None, + order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default, + direction: SQLiteDirection = SQLiteDirection.Ascending, ) -> List[AnyModelConfig]: """ Return models matching name, base and/or type. diff --git a/invokeai/app/services/model_records/model_records_sql.py b/invokeai/app/services/model_records/model_records_sql.py index edcbba2acdc..f104c3855e7 100644 --- a/invokeai/app/services/model_records/model_records_sql.py +++ b/invokeai/app/services/model_records/model_records_sql.py @@ -57,6 +57,7 @@ UnknownModelException, ) from invokeai.app.services.shared.pagination import PaginatedResults +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.services.shared.sqlite.sqlite_database import SqliteDatabase from invokeai.backend.model_manager.configs.factory import AnyModelConfig, ModelConfigFactory from invokeai.backend.model_manager.taxonomy import BaseModelType, ModelFormat, ModelType @@ -257,6 +258,7 @@ def search_by_attr( model_type: Optional[ModelType] = None, model_format: Optional[ModelFormat] = None, order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default, + direction: SQLiteDirection = SQLiteDirection.Ascending, ) -> List[AnyModelConfig]: """ Return models matching name, base and/or type. @@ -266,18 +268,24 @@ def search_by_attr( :param model_type: Filter by type of model (optional) :param model_format: Filter by model format (e.g. "diffusers") (optional) :param order_by: Result order + :param direction: Result direction If none of the optional filters are passed, will return all models in the database. """ with self._db.transaction() as cursor: assert isinstance(order_by, ModelRecordOrderBy) + order_dir = "DESC" if direction == SQLiteDirection.Descending else "ASC" ordering = { - ModelRecordOrderBy.Default: "type, base, name, format", + ModelRecordOrderBy.Default: f"type {order_dir}, base COLLATE NOCASE {order_dir}, name COLLATE NOCASE {order_dir}, format", ModelRecordOrderBy.Type: "type", - ModelRecordOrderBy.Base: "base", - ModelRecordOrderBy.Name: "name", + ModelRecordOrderBy.Base: "base COLLATE NOCASE", + ModelRecordOrderBy.Name: "name COLLATE NOCASE", ModelRecordOrderBy.Format: "format", + ModelRecordOrderBy.Size: "IFNULL(json_extract(config, '$.file_size'), 0)", + ModelRecordOrderBy.DateAdded: "created_at", + ModelRecordOrderBy.DateModified: "updated_at", + ModelRecordOrderBy.Path: "path", } where_clause: list[str] = [] @@ -301,7 +309,7 @@ def search_by_attr( SELECT config FROM models {where} - ORDER BY {ordering[order_by]} -- using ? to bind doesn't work here for some reason; + ORDER BY {ordering[order_by]} {order_dir} -- using ? to bind doesn't work here for some reason; """, tuple(bindings), ) @@ -357,17 +365,26 @@ def search_by_hash(self, hash: str) -> List[AnyModelConfig]: return results def list_models( - self, page: int = 0, per_page: int = 10, order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default + self, + page: int = 0, + per_page: int = 10, + order_by: ModelRecordOrderBy = ModelRecordOrderBy.Default, + direction: SQLiteDirection = SQLiteDirection.Ascending, ) -> PaginatedResults[ModelSummary]: """Return a paginated summary listing of each model in the database.""" with self._db.transaction() as cursor: assert isinstance(order_by, ModelRecordOrderBy) + order_dir = "DESC" if direction == SQLiteDirection.Descending else "ASC" ordering = { - ModelRecordOrderBy.Default: "type, base, name, format", + ModelRecordOrderBy.Default: f"type {order_dir}, base COLLATE NOCASE {order_dir}, name COLLATE NOCASE {order_dir}, format", ModelRecordOrderBy.Type: "type", - ModelRecordOrderBy.Base: "base", - ModelRecordOrderBy.Name: "name", + ModelRecordOrderBy.Base: "base COLLATE NOCASE", + ModelRecordOrderBy.Name: "name COLLATE NOCASE", ModelRecordOrderBy.Format: "format", + ModelRecordOrderBy.Size: "IFNULL(json_extract(config, '$.file_size'), 0)", + ModelRecordOrderBy.DateAdded: "created_at", + ModelRecordOrderBy.DateModified: "updated_at", + ModelRecordOrderBy.Path: "path", } # Lock so that the database isn't updated while we're doing the two queries. @@ -385,7 +402,7 @@ def list_models( f"""--sql SELECT config FROM models - ORDER BY {ordering[order_by]} -- using ? to bind doesn't work here for some reason + ORDER BY {ordering[order_by]} {order_dir} -- using ? to bind doesn't work here for some reason LIMIT ? OFFSET ?; """, diff --git a/invokeai/frontend/web/public/locales/en.json b/invokeai/frontend/web/public/locales/en.json index 285dc0817e1..3fce8e2861d 100644 --- a/invokeai/frontend/web/public/locales/en.json +++ b/invokeai/frontend/web/public/locales/en.json @@ -1125,6 +1125,15 @@ "modelType": "Model Type", "modelUpdated": "Model Updated", "modelUpdateFailed": "Model Update Failed", + "sortByName": "Name", + "sortByBase": "Base", + "sortBySize": "Size", + "sortByDateAdded": "Date Added", + "sortByDateModified": "Date Modified", + "sortByPath": "Path", + "sortByType": "Type", + "sortByFormat": "Format", + "sortDefault": "Default", "name": "Name", "modelPickerFallbackNoModelsInstalled": "No models installed.", "modelPickerFallbackNoModelsInstalled2": "Visit the Model Manager to install models.", diff --git a/invokeai/frontend/web/src/features/modelManagerV2/store/modelManagerV2Slice.ts b/invokeai/frontend/web/src/features/modelManagerV2/store/modelManagerV2Slice.ts index 092998d0c31..b6aea4138bb 100644 --- a/invokeai/frontend/web/src/features/modelManagerV2/store/modelManagerV2Slice.ts +++ b/invokeai/frontend/web/src/features/modelManagerV2/store/modelManagerV2Slice.ts @@ -22,6 +22,10 @@ const zModelManagerState = z.object({ scanPath: z.string().optional(), shouldInstallInPlace: z.boolean(), selectedModelKeys: z.array(z.string()), + orderBy: z + .enum(['default', 'name', 'type', 'base', 'size', 'created_at', 'updated_at', 'path', 'format']) + .default('name'), + sortDirection: z.enum(['asc', 'desc']).default('asc'), }); type ModelManagerState = z.infer; @@ -35,6 +39,8 @@ const getInitialState = (): ModelManagerState => ({ scanPath: undefined, shouldInstallInPlace: true, selectedModelKeys: [], + orderBy: 'name', + sortDirection: 'asc', }); const slice = createSlice({ @@ -74,6 +80,12 @@ const slice = createSlice({ clearModelSelection: (state) => { state.selectedModelKeys = []; }, + setOrderBy: (state, action: PayloadAction) => { + state.orderBy = action.payload; + }, + setSortDirection: (state, action: PayloadAction) => { + state.sortDirection = action.payload; + }, }, }); @@ -87,6 +99,8 @@ export const { modelSelectionChanged, toggleModelSelection, clearModelSelection, + setOrderBy, + setSortDirection, } = slice.actions; export const modelManagerSliceConfig: SliceConfig = { @@ -116,3 +130,5 @@ export const selectSearchTerm = createModelManagerSelector((mm) => mm.searchTerm export const selectFilteredModelType = createModelManagerSelector((mm) => mm.filteredModelType); export const selectShouldInstallInPlace = createModelManagerSelector((mm) => mm.shouldInstallInPlace); export const selectSelectedModelKeys = createModelManagerSelector((mm) => mm.selectedModelKeys); +export const selectOrderBy = createModelManagerSelector((mm) => mm.orderBy); +export const selectSortDirection = createModelManagerSelector((mm) => mm.sortDirection); diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelList.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelList.tsx index efb5c1add22..165150dd7f4 100644 --- a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelList.tsx +++ b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelList.tsx @@ -8,8 +8,10 @@ import { clearModelSelection, type FilterableModelType, selectFilteredModelType, + selectOrderBy, selectSearchTerm, selectSelectedModelKeys, + selectSortDirection, setSelectedModelKey, } from 'features/modelManagerV2/store/modelManagerV2Slice'; import { memo, useCallback, useMemo, useState } from 'react'; @@ -39,6 +41,8 @@ const ModelList = () => { const dispatch = useAppDispatch(); const filteredModelType = useAppSelector(selectFilteredModelType); const searchTerm = useAppSelector(selectSearchTerm); + const orderBy = useAppSelector(selectOrderBy); + const direction = useAppSelector(selectSortDirection); const selectedModelKeys = useAppSelector(selectSelectedModelKeys); const { t } = useTranslation(); const toast = useToast(); @@ -47,7 +51,8 @@ const ModelList = () => { const [isDeleting, setIsDeleting] = useState(false); const [isReidentifying, setIsReidentifying] = useState(false); - const { data: allModelsData, isLoading: isLoadingAll } = useGetModelConfigsQuery(); + const queryArgs = useMemo(() => ({ order_by: orderBy, direction: direction.toUpperCase() }), [orderBy, direction]); + const { data: allModelsData, isLoading: isLoadingAll } = useGetModelConfigsQuery(queryArgs); const { data: missingModelsData, isLoading: isLoadingMissing } = useGetMissingModelsQuery(); const [bulkDeleteModels] = useBulkDeleteModelsMutation(); const [bulkReidentifyModels] = useBulkReidentifyModelsMutation(); diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx index 78bed8ab830..2f3dcb94efe 100644 --- a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx +++ b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx @@ -7,6 +7,7 @@ import { memo, useCallback } from 'react'; import { PiXBold } from 'react-icons/pi'; import { ModelListBulkActions } from './ModelListBulkActions'; +import { ModelSortControl } from './ModelSortControl'; import { ModelTypeFilter } from './ModelTypeFilter'; export const ModelListNavigation = memo(() => { @@ -49,6 +50,9 @@ export const ModelListNavigation = memo(() => { )} + + + diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx new file mode 100644 index 00000000000..1e84de30dbf --- /dev/null +++ b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx @@ -0,0 +1,85 @@ +import { Flex, IconButton, Select } from '@invoke-ai/ui-library'; +import { useAppDispatch, useAppSelector } from 'app/store/storeHooks'; +import { + selectOrderBy, + selectSortDirection, + setOrderBy, + setSortDirection, +} from 'features/modelManagerV2/store/modelManagerV2Slice'; +import type { ChangeEvent } from 'react'; +import { useCallback, useMemo } from 'react'; +import { useTranslation } from 'react-i18next'; +import { PiSortAscendingBold, PiSortDescendingBold } from 'react-icons/pi'; +import { z } from 'zod'; + +const zOrderBy = z.enum(['default', 'name', 'type', 'base', 'size', 'created_at', 'updated_at', 'path', 'format']); +type OrderBy = z.infer; +const isOrderBy = (v: unknown): v is OrderBy => zOrderBy.safeParse(v).success; + +const ORDER_BY_OPTIONS: OrderBy[] = [ + 'default', + 'name', + 'base', + 'size', + 'created_at', + 'updated_at', + 'path', + 'type', + 'format', +]; + +export const ModelSortControl = () => { + const { t } = useTranslation(); + const dispatch = useAppDispatch(); + + const orderBy = useAppSelector(selectOrderBy); + const direction = useAppSelector(selectSortDirection); + + const ORDER_BY_LABELS = useMemo( + () => ({ + default: t('modelManager.sortDefault'), + name: t('modelManager.sortByName'), + base: t('modelManager.sortByBase'), + size: t('modelManager.sortBySize'), + created_at: t('modelManager.sortByDateAdded'), + updated_at: t('modelManager.sortByDateModified'), + path: t('modelManager.sortByPath'), + type: t('modelManager.sortByType'), + format: t('modelManager.sortByFormat'), + }), + [t] + ); + + const onChangeOrderBy = useCallback( + (e: ChangeEvent) => { + if (!isOrderBy(e.target.value)) { + return; + } + dispatch(setOrderBy(e.target.value)); + }, + [dispatch] + ); + + const toggleDirection = useCallback(() => { + dispatch(setSortDirection(direction === 'asc' ? 'desc' : 'asc')); + }, [dispatch, direction]); + + return ( + + + : } + size="sm" + variant="ghost" + onClick={toggleDirection} + /> + + ); +}; diff --git a/invokeai/frontend/web/src/services/api/endpoints/models.ts b/invokeai/frontend/web/src/services/api/endpoints/models.ts index c3d0decd53c..f279d46d823 100644 --- a/invokeai/frontend/web/src/services/api/endpoints/models.ts +++ b/invokeai/frontend/web/src/services/api/endpoints/models.ts @@ -111,9 +111,13 @@ type DeleteOrphanedModelsResponse = { errors: Record; }; +type GetModelConfigsArg = { + order_by?: string; + direction?: string; +} | void; + const modelConfigsAdapter = createEntityAdapter({ selectId: (entity) => entity.key, - sortComparer: (a, b) => a.name.localeCompare(b.name), }); export const modelConfigsAdapterSelectors = modelConfigsAdapter.getSelectors(undefined, getSelectorsOptions); @@ -338,8 +342,11 @@ export const modelsApi = api.injectEndpoints({ }, invalidatesTags: ['ModelInstalls'], }), - getModelConfigs: build.query, void>({ - query: () => ({ url: buildModelsUrl() }), + getModelConfigs: build.query, GetModelConfigsArg>({ + query: (arg) => { + const queryStr = arg ? `?${queryString.stringify(arg)}` : ''; + return { url: buildModelsUrl(queryStr) }; + }, providesTags: (result) => { const tags: ApiTagDescription[] = [{ type: 'ModelConfig', id: LIST_TAG }]; if (result) { @@ -498,5 +505,5 @@ export const { useDeleteOrphanedModelsMutation, } = modelsApi; -export const selectModelConfigsQuery = modelsApi.endpoints.getModelConfigs.select(); +export const selectModelConfigsQuery = modelsApi.endpoints.getModelConfigs.select(undefined); export const selectMissingModelsQuery = modelsApi.endpoints.getMissingModels.select(); diff --git a/tests/app/services/model_records/test_model_records_sql.py b/tests/app/services/model_records/test_model_records_sql.py index 2b6c54d5b0f..1bf697ff4cd 100644 --- a/tests/app/services/model_records/test_model_records_sql.py +++ b/tests/app/services/model_records/test_model_records_sql.py @@ -14,7 +14,9 @@ ModelRecordServiceBase, ModelRecordServiceSQL, UnknownModelException, + ModelRecordOrderBy, ) +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.services.model_records.model_records_base import ModelRecordChanges from invokeai.backend.model_manager.configs.controlnet import ControlAdapterDefaultSettings from invokeai.backend.model_manager.configs.lora import LoRA_LyCORIS_SDXL_Config @@ -363,6 +365,71 @@ def test_filter_2(store: ModelRecordServiceBase): ) assert len(matches) == 1 +def test_search_by_attr_sorting(store: ModelRecordServiceSQL): + config1 = Main_Diffusers_SD1_Config( + path="/tmp/config1", + name="alpha", + base=BaseModelType.StableDiffusion1, + type=ModelType.Main, + hash="CONFIG1HASH", + file_size=1000, + source="test/source/", + source_type=ModelSourceType.Path, + variant=ModelVariantType.Normal, + prediction_type=SchedulerPredictionType.Epsilon, + repo_variant=ModelRepoVariant.Default, + ) + config2 = Main_Diffusers_SD2_Config( + path="/tmp/config2", + name="beta", + base=BaseModelType.StableDiffusion2, + type=ModelType.Main, + hash="CONFIG2HASH", + file_size=2000, + source="test/source/", + source_type=ModelSourceType.Path, + variant=ModelVariantType.Normal, + prediction_type=SchedulerPredictionType.Epsilon, + repo_variant=ModelRepoVariant.Default, + ) + config3 = VAE_Diffusers_SD1_Config( + path="/tmp/config3", + name="gamma", + base=BaseModelType.StableDiffusion1, + type=ModelType.VAE, + hash="CONFIG3HASH", + file_size=500, + source="test/source/", + source_type=ModelSourceType.Path, + repo_variant=ModelRepoVariant.Default, + ) + for c in config1, config2, config3: + store.add_model(c) + + # Test sorting by Name Ascending + matches = store.search_by_attr(order_by=ModelRecordOrderBy.Name, direction=SQLiteDirection.Ascending) + assert len(matches) == 3 + assert matches[0].name == "alpha" + assert matches[1].name == "beta" + assert matches[2].name == "gamma" + + # Test sorting by Name Descending + matches = store.search_by_attr(order_by=ModelRecordOrderBy.Name, direction=SQLiteDirection.Descending) + assert matches[0].name == "gamma" + assert matches[1].name == "beta" + assert matches[2].name == "alpha" + + # Test sorting by Size Ascending + matches = store.search_by_attr(order_by=ModelRecordOrderBy.Size, direction=SQLiteDirection.Ascending) + assert matches[0].name == "gamma" # 500 + assert matches[1].name == "alpha" # 1000 + assert matches[2].name == "beta" # 2000 + + # Test sorting by Size Descending + matches = store.search_by_attr(order_by=ModelRecordOrderBy.Size, direction=SQLiteDirection.Descending) + assert matches[0].name == "beta" # 2000 + assert matches[1].name == "alpha" # 1000 + assert matches[2].name == "gamma" # 500 def test_model_record_changes(): # This test guards against some unexpected behaviours from pydantic's union evaluation. See #6035 From fe7bcd298eeeacacfd2a83f2b2ecbc9fe503827d Mon Sep 17 00:00:00 2001 From: skunkworxdark Date: Sun, 5 Apr 2026 21:12:48 +0100 Subject: [PATCH 3/8] ruff fix --- invokeai/app/api/routers/model_manager.py | 17 ++++++++++++++--- .../model_records/test_model_records_sql.py | 18 ++++++++++-------- 2 files changed, 24 insertions(+), 11 deletions(-) diff --git a/invokeai/app/api/routers/model_manager.py b/invokeai/app/api/routers/model_manager.py index f510cd8e382..632b5e26f0e 100644 --- a/invokeai/app/api/routers/model_manager.py +++ b/invokeai/app/api/routers/model_manager.py @@ -29,8 +29,8 @@ ModelRecordOrderBy, UnknownModelException, ) -from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.services.orphaned_models import OrphanedModelInfo +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.util.suppress_output import SuppressOutput from invokeai.backend.model_manager.configs.factory import AnyModelConfig, ModelConfigFactory from invokeai.backend.model_manager.configs.main import ( @@ -142,12 +142,23 @@ async def list_model_records( for base_model in base_models: found_models.extend( record_store.search_by_attr( - base_model=base_model, model_type=model_type, model_name=model_name, model_format=model_format, order_by=order_by, direction=direction + base_model=base_model, + model_type=model_type, + model_name=model_name, + model_format=model_format, + order_by=order_by, + direction=direction, ) ) else: found_models.extend( - record_store.search_by_attr(model_type=model_type, model_name=model_name, model_format=model_format, order_by=order_by, direction=direction) + record_store.search_by_attr( + model_type=model_type, + model_name=model_name, + model_format=model_format, + order_by=order_by, + direction=direction, + ) ) for model in found_models: model = add_cover_image_to_model_config(model, ApiDependencies) diff --git a/tests/app/services/model_records/test_model_records_sql.py b/tests/app/services/model_records/test_model_records_sql.py index 1bf697ff4cd..19a1b74e73f 100644 --- a/tests/app/services/model_records/test_model_records_sql.py +++ b/tests/app/services/model_records/test_model_records_sql.py @@ -11,13 +11,13 @@ from invokeai.app.services.config import InvokeAIAppConfig from invokeai.app.services.model_records import ( DuplicateModelException, + ModelRecordOrderBy, ModelRecordServiceBase, ModelRecordServiceSQL, UnknownModelException, - ModelRecordOrderBy, ) -from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.app.services.model_records.model_records_base import ModelRecordChanges +from invokeai.app.services.shared.sqlite.sqlite_common import SQLiteDirection from invokeai.backend.model_manager.configs.controlnet import ControlAdapterDefaultSettings from invokeai.backend.model_manager.configs.lora import LoRA_LyCORIS_SDXL_Config from invokeai.backend.model_manager.configs.main import ( @@ -365,6 +365,7 @@ def test_filter_2(store: ModelRecordServiceBase): ) assert len(matches) == 1 + def test_search_by_attr_sorting(store: ModelRecordServiceSQL): config1 = Main_Diffusers_SD1_Config( path="/tmp/config1", @@ -421,15 +422,16 @@ def test_search_by_attr_sorting(store: ModelRecordServiceSQL): # Test sorting by Size Ascending matches = store.search_by_attr(order_by=ModelRecordOrderBy.Size, direction=SQLiteDirection.Ascending) - assert matches[0].name == "gamma" # 500 - assert matches[1].name == "alpha" # 1000 - assert matches[2].name == "beta" # 2000 + assert matches[0].name == "gamma" # 500 + assert matches[1].name == "alpha" # 1000 + assert matches[2].name == "beta" # 2000 # Test sorting by Size Descending matches = store.search_by_attr(order_by=ModelRecordOrderBy.Size, direction=SQLiteDirection.Descending) - assert matches[0].name == "beta" # 2000 - assert matches[1].name == "alpha" # 1000 - assert matches[2].name == "gamma" # 500 + assert matches[0].name == "beta" # 2000 + assert matches[1].name == "alpha" # 1000 + assert matches[2].name == "gamma" # 500 + def test_model_record_changes(): # This test guards against some unexpected behaviours from pydantic's union evaluation. See #6035 From 3fe76eaf2b5bfcb5a704eab3e9956e1d5745e05f Mon Sep 17 00:00:00 2001 From: skunkworxdark Date: Sun, 5 Apr 2026 21:15:53 +0100 Subject: [PATCH 4/8] typegen fix --- .../frontend/web/src/services/api/schema.ts | 40190 +++++++++++----- 1 file changed, 27592 insertions(+), 12598 deletions(-) diff --git a/invokeai/frontend/web/src/services/api/schema.ts b/invokeai/frontend/web/src/services/api/schema.ts index 2ccb070c6fc..d2ad4c8f10d 100644 --- a/invokeai/frontend/web/src/services/api/schema.ts +++ b/invokeai/frontend/web/src/services/api/schema.ts @@ -2594,6 +2594,216 @@ export type components = { */ is_active?: boolean | null; }; + /** + * Adv AutoStereogram + * @description create an advanced autostereogram from a depth map + */ + AdvAutostereogramInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The depth map to create the image from + * @default null + */ + depth_map?: components["schemas"]["ImageField"] | null; + /** + * @description The pattern image to use as the background, if not provided then random dots will be used + * @default null + */ + pattern?: components["schemas"]["ImageField"] | null; + /** + * Pattern Width + * @description The pattern width pixels + * @default 100 + */ + pattern_width?: number; + /** + * Depth Steps + * @description The number of depth steps, 30-127 is a good range but should be less than the pattern width + * @default 50 + */ + depth_steps?: number; + /** + * Invert Depth Map + * @description Invert the depth map (difference between crossing and uncrossing eyes) + * @default false + */ + invert_depth_map?: boolean; + /** + * Grayscale + * @description Color or Grayscale output + * @default false + */ + grayscale?: boolean; + /** + * type + * @default adv_autostereogram + * @constant + */ + type: "adv_autostereogram"; + }; + /** + * Advanced Text Font to Image + * @description Overlay Text onto an image or blank canvas. + */ + AdvancedTextFontImageInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * Text Input + * @description The text from which to generate an image + * @default Invoke AI + */ + text_input?: string; + /** + * Text Input Second Row + * @description The second row of text to add below the first text + * @default null + */ + text_input_second_row?: string | null; + /** + * Font Url + * @description URL address of the font file to download + * @default https://www.1001fonts.com/download/font/caliban.medium.ttf + */ + font_url?: string | null; + /** + * Local Font Path + * @description Local font file path (overrides font_url) + * @default null + */ + local_font_path?: string | null; + /** + * Local Font + * @description Name of the local font file to use from the font_cache folder + * @default None + * @constant + */ + local_font?: "None"; + /** + * Image Width + * @description Width of the output image + * @default 1024 + */ + image_width?: number; + /** + * Image Height + * @description Height of the output image + * @default 512 + */ + image_height?: number; + /** + * Font Color First + * @description Font color for the first row of text in HEX format (e.g., '#FFFFFF') + * @default #FFFFFF + */ + font_color_first?: string; + /** + * X Position First + * @description X position of the first row of text + * @default 0 + */ + x_position_first?: number; + /** + * Y Position First + * @description Y position of the first row of text + * @default 0 + */ + y_position_first?: number; + /** + * Rotation First + * @description Rotation angle of the first row of text (in degrees) + * @default 0 + */ + rotation_first?: number; + /** + * Font Size First + * @description Font size for the first row of text + * @default 35 + */ + font_size_first?: number | null; + /** + * Font Color Second + * @description Font color for the second row of text in HEX format (e.g., '#FFFFFF') + * @default #FFFFFF + */ + font_color_second?: string; + /** + * X Position Second + * @description X position of the second row of text + * @default 0 + */ + x_position_second?: number; + /** + * Y Position Second + * @description Y position of the second row of text + * @default 0 + */ + y_position_second?: number; + /** + * Rotation Second + * @description Rotation angle of the second row of text (in degrees) + * @default 0 + */ + rotation_second?: number; + /** + * Font Size Second + * @description Font size for the second row of text + * @default 35 + */ + font_size_second?: number | null; + /** + * @description An image to place the text onto + * @default null + */ + input_image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default Advanced_Text_Font_to_Image + * @constant + */ + type: "Advanced_Text_Font_to_Image"; + }; /** * Alpha Mask to Tensor * @description Convert a mask image to a tensor. Opaque regions are 1 and transparent regions are 0. @@ -2634,6 +2844,102 @@ export type components = { */ type: "alpha_mask_to_tensor"; }; + /** + * Anamorphic Streaks + * @description Adds anamorphic streaks to the input image + */ + AnamorphicStreaksInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to streak + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description The color to use for streaks + * @default { + * "r": 0, + * "g": 100, + * "b": 255, + * "a": 255 + * } + */ + streak_color?: components["schemas"]["ColorField"]; + /** + * Gamma + * @description Gamma to use for finding highlights + * @default 30 + */ + gamma?: number; + /** + * Blur Radius + * @description Radius to blur highlights + * @default 5 + */ + blur_radius?: number; + /** + * Erosion Kernel Size + * @description Amount to erode blurred highlights + * @default 5 + */ + erosion_kernel_size?: number; + /** + * Erosion Iterations + * @description Number of erosion iterations + * @default 2 + */ + erosion_iterations?: number; + /** + * Streak Intensity + * @description Streak Intensity + * @default 10 + */ + streak_intensity?: number; + /** + * Internal Reflection Strength + * @description Internal reflection strength + * @default 0.3 + */ + internal_reflection_strength?: number; + /** + * Streak Width + * @description Streak width (recommended image width) + * @default 512 + */ + streak_width?: number; + /** + * type + * @default anamorphic_streaks + * @constant + */ + type: "anamorphic_streaks"; + }; AnyModelConfig: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; /** * AppVersion @@ -2760,63 +3066,181 @@ export type components = { type: "apply_mask_to_image"; }; /** - * BaseMetadata - * @description Adds typing data for discriminated union. + * AutoStereogram + * @description create an autostereogram from a depth map */ - BaseMetadata: { + AutostereogramInvocation: { /** - * Name - * @description model's name + * @description The board to save the image to + * @default null */ - name: string; + board?: components["schemas"]["BoardField"] | null; /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} + * @description Optional metadata to be saved with the image + * @default null */ - type: "basemetadata"; - }; - /** - * BaseModelType - * @description An enumeration of base model architectures. For example, Stable Diffusion 1.x, Stable Diffusion 2.x, FLUX, etc. - * - * Every model config must have a base architecture type. - * - * Not all models are associated with a base architecture. For example, CLIP models are their own thing, not related - * to any particular model architecture. To simplify internal APIs and make it easier to work with models, we use a - * fallback/null value `BaseModelType.Any` for these models, instead of making the model base optional. - * @enum {string} - */ - BaseModelType: "any" | "sd-1" | "sd-2" | "sd-3" | "sdxl" | "sdxl-refiner" | "flux" | "flux2" | "cogview4" | "z-image" | "unknown"; - /** Batch */ - Batch: { + metadata?: components["schemas"]["MetadataField"] | null; /** - * Batch Id - * @description The ID of the batch + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - batch_id?: string; + id: string; /** - * Origin - * @description The origin of this queue item. This data is used by the frontend to determine how to handle results. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - origin?: string | null; + is_intermediate?: boolean; /** - * Destination - * @description The origin of this queue item. This data is used by the frontend to determine how to handle results + * Use Cache + * @description Whether or not to use the cache + * @default true */ - destination?: string | null; + use_cache?: boolean; /** - * Data - * @description The batch data collection. + * @description The depth map to create the autostereogram from + * @default null */ - data?: components["schemas"]["BatchDatum"][][] | null; - /** @description The graph to initialize the session with */ - graph: components["schemas"]["Graph"]; - /** @description The workflow to initialize the session with */ - workflow?: components["schemas"]["WorkflowWithoutID"] | null; + depth_map?: components["schemas"]["ImageField"] | null; /** - * Runs - * @description Int stating how many times to iterate through all possible batch indices - * @default 1 + * @description The pattern image, if not provided then random dots will be used + * @default null + */ + pattern?: components["schemas"]["ImageField"] | null; + /** + * Pattern Divisions + * @description How many pattern repeats in output 5-10 is in general a good range. lower = more depth but harder to see + * @default 8 + */ + pattern_divisions?: number; + /** + * Invert Depth Map + * @description Invert the depth map (difference between crossing and uncrossing eyes) + * @default false + */ + invert_depth_map?: boolean; + /** + * Grayscale + * @description Color or Grayscale output + * @default false + */ + grayscale?: boolean; + /** + * type + * @default autostereogram + * @constant + */ + type: "autostereogram"; + }; + /** + * Average Images + * @description Average images + */ + AverageImagesInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Images + * @description The collection of images to average + * @default null + */ + images?: components["schemas"]["ImageField"][] | null; + /** + * Gamma + * @description Gamma for color correcting before/after blending + * @default 2.2 + */ + gamma?: number; + /** + * type + * @default average_images + * @constant + */ + type: "average_images"; + }; + /** + * BaseMetadata + * @description Adds typing data for discriminated union. + */ + BaseMetadata: { + /** + * Name + * @description model's name + */ + name: string; + /** + * @description discriminator enum property added by openapi-typescript + * @enum {string} + */ + type: "basemetadata"; + }; + /** + * BaseModelType + * @description An enumeration of base model architectures. For example, Stable Diffusion 1.x, Stable Diffusion 2.x, FLUX, etc. + * + * Every model config must have a base architecture type. + * + * Not all models are associated with a base architecture. For example, CLIP models are their own thing, not related + * to any particular model architecture. To simplify internal APIs and make it easier to work with models, we use a + * fallback/null value `BaseModelType.Any` for these models, instead of making the model base optional. + * @enum {string} + */ + BaseModelType: "any" | "sd-1" | "sd-2" | "sd-3" | "sdxl" | "sdxl-refiner" | "flux" | "flux2" | "cogview4" | "z-image" | "unknown"; + /** Batch */ + Batch: { + /** + * Batch Id + * @description The ID of the batch + */ + batch_id?: string; + /** + * Origin + * @description The origin of this queue item. This data is used by the frontend to determine how to handle results. + */ + origin?: string | null; + /** + * Destination + * @description The origin of this queue item. This data is used by the frontend to determine how to handle results + */ + destination?: string | null; + /** + * Data + * @description The batch data collection. + */ + data?: components["schemas"]["BatchDatum"][][] | null; + /** @description The graph to initialize the session with */ + graph: components["schemas"]["Graph"]; + /** @description The workflow to initialize the session with */ + workflow?: components["schemas"]["WorkflowWithoutID"] | null; + /** + * Runs + * @description Int stating how many times to iterate through all possible batch indices + * @default 1 */ runs: number; }; @@ -3418,6 +3842,147 @@ export type components = { */ metadata?: string | null; }; + /** + * Bool Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option + */ + BoolCollectionIndexInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * Random + * @description Random Index? + * @default true + */ + random?: boolean; + /** + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 + */ + index?: number; + /** + * Collection + * @description bool collection + * @default null + */ + collection?: boolean[] | null; + /** + * type + * @default bool_collection_index + * @constant + */ + type: "bool_collection_index"; + }; + /** + * Bool Collection Toggle + * @description Allows boolean selection between two separate boolean collection inputs + */ + BoolCollectionToggleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * Col1 + * @description First Bool Collection Input + * @default null + */ + col1?: boolean[] | null; + /** + * Col2 + * @description Second Bool Collection Input + * @default null + */ + col2?: boolean[] | null; + /** + * type + * @default bool_collection_toggle + * @constant + */ + type: "bool_collection_toggle"; + }; + /** + * Bool Toggle + * @description Allows boolean selection between two separate boolean inputs + */ + BoolToggleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * Bool1 + * @description First Bool Input + * @default false + */ + bool1?: boolean; + /** + * Bool2 + * @description Second Bool Input + * @default false + */ + bool2?: boolean; + /** + * type + * @default bool_toggle + * @constant + */ + type: "bool_toggle"; + }; /** * Boolean Collection Primitive * @description A collection of boolean primitive values @@ -3453,6 +4018,47 @@ export type components = { */ type: "boolean_collection"; }; + /** + * Boolean Collection Primitive Linked + * @description A collection of boolean primitive values + */ + BooleanCollectionLinkedInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Collection + * @description The collection of boolean values + * @default [] + */ + collection?: boolean[]; + /** + * type + * @default boolean_collection_linked + * @constant + */ + type: "boolean_collection_linked"; + /** + * Value + * @description The boolean value + * @default false + */ + value?: boolean; + }; /** * BooleanCollectionOutput * @description Base class for nodes that output a collection of booleans @@ -4113,10 +4719,10 @@ export type components = { cpu_only: boolean | null; }; /** - * CV2 Infill - * @description Infills transparent areas of an image using OpenCV Inpainting + * CMYK Color Separation + * @description Get color images from a base color and two others that subtractively mix to obtain it */ - CV2InfillInvocation: { + CMYKColorSeparationInvocation: { /** * @description The board to save the image to * @default null @@ -4145,59 +4751,94 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process - * @default null + * Width + * @description Desired image width + * @default 512 */ - image?: components["schemas"]["ImageField"] | null; + width?: number; /** - * type - * @default infill_cv2 - * @constant + * Height + * @description Desired image height + * @default 512 */ - type: "infill_cv2"; - }; - /** CacheStats */ - CacheStats: { + height?: number; /** - * Hits + * C Value + * @description Desired final cyan value * @default 0 */ - hits?: number; + c_value?: number; /** - * Misses - * @default 0 + * M Value + * @description Desired final magenta value + * @default 25 */ - misses?: number; + m_value?: number; /** - * High Watermark - * @default 0 + * Y Value + * @description Desired final yellow value + * @default 28 */ - high_watermark?: number; + y_value?: number; /** - * In Cache - * @default 0 + * K Value + * @description Desired final black value + * @default 76 */ - in_cache?: number; + k_value?: number; /** - * Cleared - * @default 0 + * C Split + * @description Desired cyan split point % [0..1.0] + * @default 0.5 */ - cleared?: number; + c_split?: number; /** - * Cache Size + * M Split + * @description Desired magenta split point % [0..1.0] + * @default 1 + */ + m_split?: number; + /** + * Y Split + * @description Desired yellow split point % [0..1.0] * @default 0 */ - cache_size?: number; - /** Loaded Model Sizes */ - loaded_model_sizes?: { - [key: string]: number; - }; + y_split?: number; + /** + * K Split + * @description Desired black split point % [0..1.0] + * @default 0.5 + */ + k_split?: number; + /** + * Profile + * @description CMYK Color Profile + * @default Default + * @enum {string} + */ + profile?: "Default" | "PIL"; + /** + * type + * @default cmyk_separation + * @constant + */ + type: "cmyk_separation"; }; /** - * Calculate Image Tiles Even Split - * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. + * CMYK Halftone + * @description Halftones an image in the style of a CMYK print */ - CalculateImageTilesEvenSplitInvocation: { + CMYKHalftoneInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4216,106 +4857,92 @@ export type components = { */ use_cache?: boolean; /** - * Image Width - * @description The image width, in pixels, to calculate tiles for. - * @default 1024 + * @description The image to halftone + * @default null */ - image_width?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Image Height - * @description The image height, in pixels, to calculate tiles for. - * @default 1024 + * Spacing + * @description Halftone dot spacing + * @default 8 */ - image_height?: number; + spacing?: number; /** - * Num Tiles X - * @description Number of tiles to divide image into on the x axis - * @default 2 + * C Angle + * @description C halftone angle + * @default 15 */ - num_tiles_x?: number; + c_angle?: number; /** - * Num Tiles Y - * @description Number of tiles to divide image into on the y axis - * @default 2 + * M Angle + * @description M halftone angle + * @default 75 */ - num_tiles_y?: number; + m_angle?: number; /** - * Overlap - * @description The overlap, in pixels, between adjacent tiles. - * @default 128 + * Y Angle + * @description Y halftone angle + * @default 90 */ - overlap?: number; + y_angle?: number; /** - * type - * @default calculate_image_tiles_even_split - * @constant + * K Angle + * @description K halftone angle + * @default 45 */ - type: "calculate_image_tiles_even_split"; - }; - /** - * Calculate Image Tiles - * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. - */ - CalculateImageTilesInvocation: { + k_angle?: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Oversampling + * @description Oversampling factor + * @default 1 */ - id: string; + oversampling?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Offset C + * @description Offset Cyan halfway between dots * @default false */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Image Width - * @description The image width, in pixels, to calculate tiles for. - * @default 1024 - */ - image_width?: number; - /** - * Image Height - * @description The image height, in pixels, to calculate tiles for. - * @default 1024 - */ - image_height?: number; + offset_c?: boolean; /** - * Tile Width - * @description The tile width, in pixels. - * @default 576 + * Offset M + * @description Offset Magenta halfway between dots + * @default false */ - tile_width?: number; + offset_m?: boolean; /** - * Tile Height - * @description The tile height, in pixels. - * @default 576 + * Offset Y + * @description Offset Yellow halfway between dots + * @default false */ - tile_height?: number; + offset_y?: boolean; /** - * Overlap - * @description The target overlap, in pixels, between adjacent tiles. Adjacent tiles will overlap by at least this amount - * @default 128 + * Offset K + * @description Offset K halfway between dots + * @default false */ - overlap?: number; + offset_k?: boolean; /** * type - * @default calculate_image_tiles + * @default cmyk_halftone * @constant */ - type: "calculate_image_tiles"; + type: "cmyk_halftone"; }; /** - * Calculate Image Tiles Minimum Overlap - * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. + * CMYK Merge + * @description Merge subtractive color channels (CMYK+alpha) */ - CalculateImageTilesMinimumOverlapInvocation: { + CMYKMergeInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4334,94 +4961,107 @@ export type components = { */ use_cache?: boolean; /** - * Image Width - * @description The image width, in pixels, to calculate tiles for. - * @default 1024 - */ - image_width?: number; - /** - * Image Height - * @description The image height, in pixels, to calculate tiles for. - * @default 1024 + * @description The c channel + * @default null */ - image_height?: number; + c_channel?: components["schemas"]["ImageField"] | null; /** - * Tile Width - * @description The tile width, in pixels. - * @default 576 + * @description The m channel + * @default null */ - tile_width?: number; + m_channel?: components["schemas"]["ImageField"] | null; /** - * Tile Height - * @description The tile height, in pixels. - * @default 576 + * @description The y channel + * @default null */ - tile_height?: number; + y_channel?: components["schemas"]["ImageField"] | null; /** - * Min Overlap - * @description Minimum overlap between adjacent tiles, in pixels. - * @default 128 + * @description The k channel + * @default null */ - min_overlap?: number; + k_channel?: components["schemas"]["ImageField"] | null; /** - * type - * @default calculate_image_tiles_min_overlap - * @constant + * @description The alpha channel + * @default null */ - type: "calculate_image_tiles_min_overlap"; - }; - /** CalculateImageTilesOutput */ - CalculateImageTilesOutput: { + alpha_channel?: components["schemas"]["ImageField"] | null; /** - * Tiles - * @description The tiles coordinates that cover a particular image shape. + * Profile + * @description CMYK Color Profile + * @default Default + * @enum {string} */ - tiles: components["schemas"]["Tile"][]; + profile?: "Default" | "PIL"; /** * type - * @default calculate_image_tiles_output + * @default cmyk_merge * @constant */ - type: "calculate_image_tiles_output"; + type: "cmyk_merge"; }; /** - * CancelAllExceptCurrentResult - * @description Result of canceling all except current + * CMYKSeparationOutput + * @description Base class for invocations that output four L-mode images (C, M, Y, K) */ - CancelAllExceptCurrentResult: { + CMYKSeparationOutput: { + /** @description Blank image of the specified color */ + color_image: components["schemas"]["ImageField"]; /** - * Canceled - * @description Number of queue items canceled + * Width + * @description The width of the image in pixels */ - canceled: number; - }; - /** - * CancelByBatchIDsResult - * @description Result of canceling by list of batch ids - */ - CancelByBatchIDsResult: { + width: number; /** - * Canceled - * @description Number of queue items canceled + * Height + * @description The height of the image in pixels */ - canceled: number; - }; - /** - * CancelByDestinationResult - * @description Result of canceling by a destination - */ - CancelByDestinationResult: { + height: number; + /** @description Blank image of the first separated color */ + part_a: components["schemas"]["ImageField"]; /** - * Canceled - * @description Number of queue items canceled + * Rgb Red A + * @description R value of color part A */ - canceled: number; + rgb_red_a: number; + /** + * Rgb Green A + * @description G value of color part A + */ + rgb_green_a: number; + /** + * Rgb Blue A + * @description B value of color part A + */ + rgb_blue_a: number; + /** @description Blank image of the second separated color */ + part_b: components["schemas"]["ImageField"]; + /** + * Rgb Red B + * @description R value of color part B + */ + rgb_red_b: number; + /** + * Rgb Green B + * @description G value of color part B + */ + rgb_green_b: number; + /** + * Rgb Blue B + * @description B value of color part B + */ + rgb_blue_b: number; + /** + * type + * @default cmyk_separation_output + * @constant + */ + type: "cmyk_separation_output"; }; /** - * Canny Edge Detection - * @description Geneartes an edge map using a cv2's Canny algorithm. + * CMYK Split + * @description Split an image into subtractive color channels (CMYK+alpha) */ - CannyEdgeDetectionInvocation: { + CMYKSplitInvocation: { /** * @description The board to save the image to * @default null @@ -4450,44 +5090,61 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The image to split into additive channels * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Low Threshold - * @description The low threshold of the Canny pixel gradient (0-255) - * @default 100 - */ - low_threshold?: number; - /** - * High Threshold - * @description The high threshold of the Canny pixel gradient (0-255) - * @default 200 + * Profile + * @description CMYK Color Profile + * @default Default + * @enum {string} */ - high_threshold?: number; + profile?: "Default" | "PIL"; /** * type - * @default canny_edge_detection + * @default cmyk_split * @constant */ - type: "canny_edge_detection"; + type: "cmyk_split"; }; /** - * Canvas Paste Back - * @description Combines two images by using the mask provided. Intended for use on the Unified Canvas. + * CMYKSplitOutput + * @description Base class for invocations that output four L-mode images (C, M, Y, K) */ - CanvasPasteBackInvocation: { + CMYKSplitOutput: { + /** @description Grayscale image of the cyan channel */ + c_channel: components["schemas"]["ImageField"]; + /** @description Grayscale image of the magenta channel */ + m_channel: components["schemas"]["ImageField"]; + /** @description Grayscale image of the yellow channel */ + y_channel: components["schemas"]["ImageField"]; + /** @description Grayscale image of the k channel */ + k_channel: components["schemas"]["ImageField"]; + /** @description Grayscale image of the alpha channel */ + alpha_channel: components["schemas"]["ImageField"]; /** - * @description The board to save the image to - * @default null + * Width + * @description The width of the image in pixels */ - board?: components["schemas"]["BoardField"] | null; + width: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Height + * @description The height of the image in pixels */ - metadata?: components["schemas"]["MetadataField"] | null; + height: number; + /** + * type + * @default cmyk_split_output + * @constant + */ + type: "cmyk_split_output"; + }; + /** + * CSV To Index String + * @description CSVToIndexString converts a CSV to a String at index with a random option + */ + CSVToIndexStringInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4502,52 +5159,39 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * @description The source image - * @default null - */ - source_image?: components["schemas"]["ImageField"] | null; - /** - * @description The target image - * @default null + * Csv String + * @description csv string + * @default */ - target_image?: components["schemas"]["ImageField"] | null; + csv_string?: string; /** - * @description The mask to use when pasting - * @default null + * Random + * @description Random Index? + * @default true */ - mask?: components["schemas"]["ImageField"] | null; + random?: boolean; /** - * Mask Blur - * @description The amount to blur the mask by + * Index + * @description zero based index into CSV array (note index will wrap around if out of bounds) * @default 0 */ - mask_blur?: number; + index?: number; /** * type - * @default canvas_paste_back + * @default csv_to_index_string * @constant */ - type: "canvas_paste_back"; + type: "csv_to_index_string"; }; /** - * Canvas V2 Mask and Crop - * @description Handles Canvas V2 image output masking and cropping + * CSV To Strings + * @description Converts a CSV string to a collection of strings */ - CanvasV2MaskAndCropInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + CSVToStringsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4566,38 +5210,33 @@ export type components = { */ use_cache?: boolean; /** - * @description The source image onto which the masked generated image is pasted. If omitted, the masked generated image is returned with transparency. - * @default null - */ - source_image?: components["schemas"]["ImageField"] | null; - /** - * @description The image to apply the mask to - * @default null - */ - generated_image?: components["schemas"]["ImageField"] | null; - /** - * @description The mask to apply + * Csv String + * @description csv string * @default null */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Mask Blur - * @description The amount to blur the mask by - * @default 0 - */ - mask_blur?: number; + csv_string?: string | null; /** * type - * @default canvas_v2_mask_and_crop + * @default csv_to_strings * @constant */ - type: "canvas_v2_mask_and_crop"; + type: "csv_to_strings"; }; /** - * Center Pad or Crop Image - * @description Pad or crop an image's sides from the center by specified pixels. Positive values are outside of the image. + * CV2 Infill + * @description Infills transparent areas of an image using OpenCV Inpainting */ - CenterPadCropInvocation: { + CV2InfillInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4616,110 +5255,59 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to crop + * @description The image to process * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Left - * @description Number of pixels to pad/crop from the left (negative values crop inwards, positive values pad outwards) - * @default 0 + * type + * @default infill_cv2 + * @constant */ - left?: number; + type: "infill_cv2"; + }; + /** CacheStats */ + CacheStats: { /** - * Right - * @description Number of pixels to pad/crop from the right (negative values crop inwards, positive values pad outwards) + * Hits * @default 0 */ - right?: number; + hits?: number; /** - * Top - * @description Number of pixels to pad/crop from the top (negative values crop inwards, positive values pad outwards) + * Misses * @default 0 */ - top?: number; + misses?: number; /** - * Bottom - * @description Number of pixels to pad/crop from the bottom (negative values crop inwards, positive values pad outwards) + * High Watermark * @default 0 */ - bottom?: number; - /** - * type - * @default img_pad_crop - * @constant - */ - type: "img_pad_crop"; - }; - /** - * Classification - * @description The classification of an Invocation. - * - `Stable`: The invocation, including its inputs/outputs and internal logic, is stable. You may build workflows with it, having confidence that they will not break because of a change in this invocation. - * - `Beta`: The invocation is not yet stable, but is planned to be stable in the future. Workflows built around this invocation may break, but we are committed to supporting this invocation long-term. - * - `Prototype`: The invocation is not yet stable and may be removed from the application at any time. Workflows built around this invocation may break, and we are *not* committed to supporting this invocation. - * - `Deprecated`: The invocation is deprecated and may be removed in a future version. - * - `Internal`: The invocation is not intended for use by end-users. It may be changed or removed at any time, but is exposed for users to play with. - * - `Special`: The invocation is a special case and does not fit into any of the other classifications. - * @enum {string} - */ - Classification: "stable" | "beta" | "prototype" | "deprecated" | "internal" | "special"; - /** - * ClearResult - * @description Result of clearing the session queue - */ - ClearResult: { + high_watermark?: number; /** - * Deleted - * @description Number of queue items deleted + * In Cache + * @default 0 */ - deleted: number; - }; - /** - * ClipVariantType - * @description Variant type. - * @enum {string} - */ - ClipVariantType: "large" | "gigantic"; - /** - * CogView4ConditioningField - * @description A conditioning tensor primitive value - */ - CogView4ConditioningField: { + in_cache?: number; /** - * Conditioning Name - * @description The name of conditioning tensor + * Cleared + * @default 0 */ - conditioning_name: string; - }; - /** - * CogView4ConditioningOutput - * @description Base class for nodes that output a CogView text conditioning tensor. - */ - CogView4ConditioningOutput: { - /** @description Conditioning tensor */ - conditioning: components["schemas"]["CogView4ConditioningField"]; + cleared?: number; /** - * type - * @default cogview4_conditioning_output - * @constant + * Cache Size + * @default 0 */ - type: "cogview4_conditioning_output"; + cache_size?: number; + /** Loaded Model Sizes */ + loaded_model_sizes?: { + [key: string]: number; + }; }; /** - * Denoise - CogView4 - * @description Run the denoising process with a CogView4 model. + * Calculate Image Tiles Even Split + * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. */ - CogView4DenoiseInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + CalculateImageTilesEvenSplitInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4738,95 +5326,106 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null + * Image Width + * @description The image width, in pixels, to calculate tiles for. + * @default 1024 */ - latents?: components["schemas"]["LatentsField"] | null; + image_width?: number; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. - * @default null + * Image Height + * @description The image height, in pixels, to calculate tiles for. + * @default 1024 */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + image_height?: number; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Num Tiles X + * @description Number of tiles to divide image into on the x axis + * @default 2 */ - denoising_start?: number; + num_tiles_x?: number; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * Num Tiles Y + * @description Number of tiles to divide image into on the y axis + * @default 2 */ - denoising_end?: number; + num_tiles_y?: number; /** - * Transformer - * @description CogView4 model (Transformer) to load - * @default null + * Overlap + * @description The overlap, in pixels, between adjacent tiles. + * @default 128 */ - transformer?: components["schemas"]["TransformerField"] | null; + overlap?: number; /** - * @description Positive conditioning tensor - * @default null + * type + * @default calculate_image_tiles_even_split + * @constant */ - positive_conditioning?: components["schemas"]["CogView4ConditioningField"] | null; + type: "calculate_image_tiles_even_split"; + }; + /** + * Calculate Image Tiles + * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. + */ + CalculateImageTilesInvocation: { /** - * @description Negative conditioning tensor - * @default null - */ - negative_conditioning?: components["schemas"]["CogView4ConditioningField"] | null; + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 3.5 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cfg_scale?: number | number[]; + is_intermediate?: boolean; /** - * Width - * @description Width of the generated image. + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Image Width + * @description The image width, in pixels, to calculate tiles for. * @default 1024 */ - width?: number; + image_width?: number; /** - * Height - * @description Height of the generated image. + * Image Height + * @description The image height, in pixels, to calculate tiles for. * @default 1024 */ - height?: number; + image_height?: number; /** - * Steps - * @description Number of steps to run - * @default 25 + * Tile Width + * @description The tile width, in pixels. + * @default 576 */ - steps?: number; + tile_width?: number; /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * Tile Height + * @description The tile height, in pixels. + * @default 576 */ - seed?: number; + tile_height?: number; + /** + * Overlap + * @description The target overlap, in pixels, between adjacent tiles. Adjacent tiles will overlap by at least this amount + * @default 128 + */ + overlap?: number; /** * type - * @default cogview4_denoise + * @default calculate_image_tiles * @constant */ - type: "cogview4_denoise"; + type: "calculate_image_tiles"; }; /** - * Image to Latents - CogView4 - * @description Generates latents from an image. + * Calculate Image Tiles Minimum Overlap + * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. */ - CogView4ImageToLatentsInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + CalculateImageTilesMinimumOverlapInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4845,27 +5444,94 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to encode. - * @default null + * Image Width + * @description The image width, in pixels, to calculate tiles for. + * @default 1024 */ - image?: components["schemas"]["ImageField"] | null; + image_width?: number; /** - * @description VAE - * @default null + * Image Height + * @description The image height, in pixels, to calculate tiles for. + * @default 1024 */ - vae?: components["schemas"]["VAEField"] | null; + image_height?: number; + /** + * Tile Width + * @description The tile width, in pixels. + * @default 576 + */ + tile_width?: number; + /** + * Tile Height + * @description The tile height, in pixels. + * @default 576 + */ + tile_height?: number; + /** + * Min Overlap + * @description Minimum overlap between adjacent tiles, in pixels. + * @default 128 + */ + min_overlap?: number; /** * type - * @default cogview4_i2l + * @default calculate_image_tiles_min_overlap * @constant */ - type: "cogview4_i2l"; + type: "calculate_image_tiles_min_overlap"; + }; + /** CalculateImageTilesOutput */ + CalculateImageTilesOutput: { + /** + * Tiles + * @description The tiles coordinates that cover a particular image shape. + */ + tiles: components["schemas"]["Tile"][]; + /** + * type + * @default calculate_image_tiles_output + * @constant + */ + type: "calculate_image_tiles_output"; }; /** - * Latents to Image - CogView4 - * @description Generates an image from latents. + * CancelAllExceptCurrentResult + * @description Result of canceling all except current */ - CogView4LatentsToImageInvocation: { + CancelAllExceptCurrentResult: { + /** + * Canceled + * @description Number of queue items canceled + */ + canceled: number; + }; + /** + * CancelByBatchIDsResult + * @description Result of canceling by list of batch ids + */ + CancelByBatchIDsResult: { + /** + * Canceled + * @description Number of queue items canceled + */ + canceled: number; + }; + /** + * CancelByDestinationResult + * @description Result of canceling by a destination + */ + CancelByDestinationResult: { + /** + * Canceled + * @description Number of queue items canceled + */ + canceled: number; + }; + /** + * Canny Edge Detection + * @description Geneartes an edge map using a cv2's Canny algorithm. + */ + CannyEdgeDetectionInvocation: { /** * @description The board to save the image to * @default null @@ -4894,27 +5560,44 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * @description The image to process * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description VAE - * @default null + * Low Threshold + * @description The low threshold of the Canny pixel gradient (0-255) + * @default 100 */ - vae?: components["schemas"]["VAEField"] | null; + low_threshold?: number; + /** + * High Threshold + * @description The high threshold of the Canny pixel gradient (0-255) + * @default 200 + */ + high_threshold?: number; /** * type - * @default cogview4_l2i + * @default canny_edge_detection * @constant */ - type: "cogview4_l2i"; + type: "canny_edge_detection"; }; /** - * Main Model - CogView4 - * @description Loads a CogView4 base model, outputting its submodels. + * Canvas Paste Back + * @description Combines two images by using the mask provided. Intended for use on the Unified Canvas. */ - CogView4ModelLoaderInvocation: { + CanvasPasteBackInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4932,47 +5615,49 @@ export type components = { * @default true */ use_cache?: boolean; - /** @description CogView4 model (Transformer) to load */ - model: components["schemas"]["ModelIdentifierField"]; /** - * type - * @default cogview4_model_loader - * @constant + * @description The source image + * @default null */ - type: "cogview4_model_loader"; - }; - /** - * CogView4ModelLoaderOutput - * @description CogView4 base model loader output. - */ - CogView4ModelLoaderOutput: { + source_image?: components["schemas"]["ImageField"] | null; /** - * Transformer - * @description Transformer + * @description The target image + * @default null */ - transformer: components["schemas"]["TransformerField"]; + target_image?: components["schemas"]["ImageField"] | null; /** - * GLM Encoder - * @description GLM (THUDM) tokenizer and text encoder + * @description The mask to use when pasting + * @default null */ - glm_encoder: components["schemas"]["GlmEncoderField"]; + mask?: components["schemas"]["ImageField"] | null; /** - * VAE - * @description VAE + * Mask Blur + * @description The amount to blur the mask by + * @default 0 */ - vae: components["schemas"]["VAEField"]; + mask_blur?: number; /** * type - * @default cogview4_model_loader_output + * @default canvas_paste_back * @constant */ - type: "cogview4_model_loader_output"; + type: "canvas_paste_back"; }; /** - * Prompt - CogView4 - * @description Encodes and preps a prompt for a cogview4 image. + * Canvas V2 Mask and Crop + * @description Handles Canvas V2 image output masking and cropping */ - CogView4TextEncoderInvocation: { + CanvasV2MaskAndCropInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4991,29 +5676,38 @@ export type components = { */ use_cache?: boolean; /** - * Prompt - * @description Text prompt to encode. + * @description The source image onto which the masked generated image is pasted. If omitted, the masked generated image is returned with transparency. * @default null */ - prompt?: string | null; + source_image?: components["schemas"]["ImageField"] | null; /** - * GLM Encoder - * @description GLM (THUDM) tokenizer and text encoder + * @description The image to apply the mask to * @default null */ - glm_encoder?: components["schemas"]["GlmEncoderField"] | null; + generated_image?: components["schemas"]["ImageField"] | null; + /** + * @description The mask to apply + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * Mask Blur + * @description The amount to blur the mask by + * @default 0 + */ + mask_blur?: number; /** * type - * @default cogview4_text_encoder + * @default canvas_v2_mask_and_crop * @constant */ - type: "cogview4_text_encoder"; + type: "canvas_v2_mask_and_crop"; }; /** - * CollectInvocation - * @description Collects values into a collection + * Center Pad or Crop Image + * @description Pad or crop an image's sides from the center by specified pixels. Positive values are outside of the image. */ - CollectInvocation: { + CenterPadCropInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5032,61 +5726,46 @@ export type components = { */ use_cache?: boolean; /** - * Collection Item - * @description The item to collect (all inputs must be of the same type) + * @description The image to crop * @default null */ - item?: unknown | null; - /** - * Collection - * @description An optional collection to append to - * @default [] - */ - collection?: unknown[]; + image?: components["schemas"]["ImageField"] | null; /** - * type - * @default collect - * @constant + * Left + * @description Number of pixels to pad/crop from the left (negative values crop inwards, positive values pad outwards) + * @default 0 */ - type: "collect"; - }; - /** CollectInvocationOutput */ - CollectInvocationOutput: { + left?: number; /** - * Collection - * @description The collection of input items + * Right + * @description Number of pixels to pad/crop from the right (negative values crop inwards, positive values pad outwards) + * @default 0 */ - collection: unknown[]; + right?: number; /** - * type - * @default collect_output - * @constant + * Top + * @description Number of pixels to pad/crop from the top (negative values crop inwards, positive values pad outwards) + * @default 0 */ - type: "collect_output"; - }; - /** - * ColorCollectionOutput - * @description Base class for nodes that output a collection of colors - */ - ColorCollectionOutput: { + top?: number; /** - * Collection - * @description The output colors + * Bottom + * @description Number of pixels to pad/crop from the bottom (negative values crop inwards, positive values pad outwards) + * @default 0 */ - collection: components["schemas"]["ColorField"][]; + bottom?: number; /** * type - * @default color_collection_output + * @default img_pad_crop * @constant */ - type: "color_collection_output"; + type: "img_pad_crop"; }; /** - * Color Correct - * @description Matches the color histogram of a base image to a reference image, optionally - * using a mask to only color-correct certain regions of the base image. + * ChromaNoise + * @description Adds chroma-only noise (in OKLab space) to an image. */ - ColorCorrectInvocation: { + ChromaNoiseInvocation: { /** * @description The board to save the image to * @default null @@ -5115,65 +5794,68 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to color-correct + * @description The image to add chroma noise to * @default null */ - base_image?: components["schemas"]["ImageField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description Reference image for color-correction + * Amount 1 + * @description Amount of the first chroma noise layer + * @default 100 + */ + amount_1?: number; + /** + * Amount 2 + * @description Amount of the second chroma noise layer + * @default 50 + */ + amount_2?: number; + /** + * Seed 1 + * @description The first seed to use (omit for random) * @default null */ - color_reference?: components["schemas"]["ImageField"] | null; + seed_1?: number | null; /** - * @description Optional mask to limit color correction area + * Seed 2 + * @description The second seed to use (omit for random) * @default null */ - mask?: components["schemas"]["ImageField"] | null; + seed_2?: number | null; /** - * Color Space - * @description Colorspace in which to apply histogram matching - * @default RGB - * @enum {string} + * Blur 1 + * @description The strength of the first noise blur + * @default 0.5 */ - colorspace?: "RGB" | "YCbCr" | "YCbCr-Chroma" | "YCbCr-Luma"; + blur_1?: number; + /** + * Blur 2 + * @description The strength of the second noise blur + * @default 0.5 + */ + blur_2?: number; /** * type - * @default color_correct + * @default chroma_noise * @constant */ - type: "color_correct"; + type: "chroma_noise"; }; /** - * ColorField - * @description A color primitive field + * Chromatic Aberration + * @description Simulate realistic chromatic aberration in an image with controllable strength and center point. */ - ColorField: { - /** - * R - * @description The red component - */ - r: number; - /** - * G - * @description The green component - */ - g: number; + ChromaticAberrationInvocation: { /** - * B - * @description The blue component + * @description Optional metadata to be saved with the image + * @default null */ - b: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * A - * @description The alpha component + * @description The board to save the image to + * @default null */ - a: number; - }; - /** - * Color Primitive - * @description A color primitive value - */ - ColorInvocation: { + board?: components["schemas"]["BoardField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5192,27 +5874,71 @@ export type components = { */ use_cache?: boolean; /** - * @description The color value - * @default { - * "r": 0, - * "g": 0, - * "b": 0, - * "a": 255 - * } + * @description The image to apply chromatic aberration to + * @default null */ - color?: components["schemas"]["ColorField"]; + image?: components["schemas"]["ImageField"] | null; + /** + * Strength + * @description Strength of the chromatic shift (0–10) + * @default 1 + */ + strength?: number; + /** + * Center X + * @description X coordinate of center (0–1) + * @default 0.5 + */ + center_x?: number; + /** + * Center Y + * @description Y coordinate of center (0–1) + * @default 0.5 + */ + center_y?: number; /** * type - * @default color + * @default chromatic_aberration * @constant */ - type: "color"; + type: "chromatic_aberration"; }; /** - * Color Map - * @description Generates a color map from the provided image. + * Classification + * @description The classification of an Invocation. + * - `Stable`: The invocation, including its inputs/outputs and internal logic, is stable. You may build workflows with it, having confidence that they will not break because of a change in this invocation. + * - `Beta`: The invocation is not yet stable, but is planned to be stable in the future. Workflows built around this invocation may break, but we are committed to supporting this invocation long-term. + * - `Prototype`: The invocation is not yet stable and may be removed from the application at any time. Workflows built around this invocation may break, and we are *not* committed to supporting this invocation. + * - `Deprecated`: The invocation is deprecated and may be removed in a future version. + * - `Internal`: The invocation is not intended for use by end-users. It may be changed or removed at any time, but is exposed for users to play with. + * - `Special`: The invocation is a special case and does not fit into any of the other classifications. + * @enum {string} */ - ColorMapInvocation: { + Classification: "stable" | "beta" | "prototype" | "deprecated" | "internal" | "special"; + /** + * ClearResult + * @description Result of clearing the session queue + */ + ClearResult: { + /** + * Deleted + * @description Number of queue items deleted + */ + deleted: number; + }; + /** + * ClipVariantType + * @description Variant type. + * @enum {string} + */ + ClipVariantType: "large" | "gigantic"; + /** + * Clipseg Mask Hierarchy + * @description Creates a segmentation hierarchy of mutually exclusive masks from clipseg text prompts. + * + * This node takes up to seven pairs of prompts/threshold values, then descends through them hierarchically creating mutually exclusive masks out of whatever it can match from the input image. This means whatever is matched in prompt 1 will be subtracted from the match area for prompt 2; both areas will be omitted from the match area of prompt 3; etc. The idea is that by starting with foreground objects and working your way back through a scene, you can create a more-or-less complete segmentation map for the image whose constituent segments can be passed off to different masks for regional conditioning or other processing. + */ + ClipsegMaskHierarchyInvocation: { /** * @description The board to save the image to * @default null @@ -5241,156 +5967,181 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The image from which to create masks * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Tile Size - * @description Tile size - * @default 64 + * Invert Output + * @description Off: white on black / On: black on white + * @default true */ - tile_size?: number; + invert_output?: boolean; /** - * type - * @default color_map - * @constant + * Smoothing + * @description Radius of blur to apply before thresholding + * @default 4 */ - type: "color_map"; - }; - /** - * ColorOutput - * @description Base class for nodes that output a single color - */ - ColorOutput: { - /** @description The output color */ - color: components["schemas"]["ColorField"]; + smoothing?: number; /** - * type - * @default color_output - * @constant + * Prompt 1 + * @description Text to mask prompt with highest segmentation priority + * @default null */ - type: "color_output"; - }; - /** - * Prompt - SD1.5 - * @description Parse prompt using compel package to conditioning. - */ - CompelInvocation: { + prompt_1?: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Threshold 1 + * @description Detection confidence threshold for prompt 1 + * @default 0.4 */ - id: string; + threshold_1?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Prompt 2 + * @description Text to mask prompt, behind prompt 1 + * @default null */ - is_intermediate?: boolean; + prompt_2?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Threshold 2 + * @description Detection confidence threshold for prompt 2 + * @default 0.4 */ - use_cache?: boolean; + threshold_2?: number; /** - * Prompt - * @description Prompt to be parsed by Compel to create a conditioning tensor - * @default + * Prompt 3 + * @description Text to mask prompt, behind prompts 1 & 2 + * @default null */ - prompt?: string; + prompt_3?: string | null; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Threshold 3 + * @description Detection confidence threshold for prompt 3 + * @default 0.4 */ - clip?: components["schemas"]["CLIPField"] | null; + threshold_3?: number; /** - * @description A mask defining the region that this conditioning prompt applies to. + * Prompt 4 + * @description Text to mask prompt, behind prompts 1, 2, & 3 * @default null */ - mask?: components["schemas"]["TensorField"] | null; + prompt_4?: string | null; /** - * type - * @default compel - * @constant + * Threshold 4 + * @description Detection confidence threshold for prompt 4 + * @default 0.4 */ - type: "compel"; - }; - /** - * Conditioning Collection Primitive - * @description A collection of conditioning tensor primitive values - */ - ConditioningCollectionInvocation: { + threshold_4?: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Prompt 5 + * @description Text to mask prompt, behind prompts 1 thru 4 + * @default null */ - id: string; + prompt_5?: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Threshold 5 + * @description Detection confidence threshold for prompt 5 + * @default 0.4 */ - is_intermediate?: boolean; + threshold_5?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Prompt 6 + * @description Text to mask prompt, behind prompts 1 thru 5 + * @default null */ - use_cache?: boolean; + prompt_6?: string | null; /** - * Collection - * @description The collection of conditioning tensors - * @default [] + * Threshold 6 + * @description Detection confidence threshold for prompt 6 + * @default 0.4 */ - collection?: components["schemas"]["ConditioningField"][]; + threshold_6?: number; + /** + * Prompt 7 + * @description Text to mask prompt, lowest priority behind all others + * @default null + */ + prompt_7?: string | null; + /** + * Threshold 7 + * @description Detection confidence threshold for prompt 7 + * @default 0.4 + */ + threshold_7?: number; /** * type - * @default conditioning_collection + * @default clipseg_mask_hierarchy * @constant */ - type: "conditioning_collection"; + type: "clipseg_mask_hierarchy"; }; /** - * ConditioningCollectionOutput - * @description Base class for nodes that output a collection of conditioning tensors + * ClipsegMaskHierarchyOutput + * @description Class for invocations that output a hierarchy of masks */ - ConditioningCollectionOutput: { - /** - * Collection - * @description The output conditioning tensors - */ - collection: components["schemas"]["ConditioningField"][]; + ClipsegMaskHierarchyOutput: { + /** @description Mask corresponding to prompt 1 (full coverage) */ + mask_1: components["schemas"]["ImageField"]; + /** @description Mask corresponding to prompt 2 (minus mask 1) */ + mask_2: components["schemas"]["ImageField"]; + /** @description Mask corresponding to prompt 3 (minus masks 1 & 2) */ + mask_3: components["schemas"]["ImageField"]; + /** @description Mask corresponding to prompt 4 (minus masks 1, 2, & 3) */ + mask_4: components["schemas"]["ImageField"]; + /** @description Mask corresponding to prompt 5 (minus masks 1 thru 4) */ + mask_5: components["schemas"]["ImageField"]; + /** @description Mask corresponding to prompt 6 (minus masks 1 thru 5) */ + mask_6: components["schemas"]["ImageField"]; + /** @description Mask corresponding to prompt 7 (minus masks 1 thru 6) */ + mask_7: components["schemas"]["ImageField"]; + /** @description Mask coresponding to remaining unmatched image areas. */ + ground_mask: components["schemas"]["ImageField"]; /** * type - * @default conditioning_collection_output + * @default clipseg_mask_hierarchy_output * @constant */ - type: "conditioning_collection_output"; + type: "clipseg_mask_hierarchy_output"; }; /** - * ConditioningField + * CogView4ConditioningField * @description A conditioning tensor primitive value */ - ConditioningField: { + CogView4ConditioningField: { /** * Conditioning Name * @description The name of conditioning tensor */ conditioning_name: string; + }; + /** + * CogView4ConditioningOutput + * @description Base class for nodes that output a CogView text conditioning tensor. + */ + CogView4ConditioningOutput: { + /** @description Conditioning tensor */ + conditioning: components["schemas"]["CogView4ConditioningField"]; /** - * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. - * @default null + * type + * @default cogview4_conditioning_output + * @constant */ - mask?: components["schemas"]["TensorField"] | null; + type: "cogview4_conditioning_output"; }; /** - * Conditioning Primitive - * @description A conditioning tensor primitive value + * Denoise - CogView4 + * @description Run the denoising process with a CogView4 model. */ - ConditioningInvocation: { + CogView4DenoiseInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5409,36 +6160,85 @@ export type components = { */ use_cache?: boolean; /** - * @description Conditioning tensor + * @description Latents tensor * @default null */ - conditioning?: components["schemas"]["ConditioningField"] | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * type - * @default conditioning - * @constant + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - type: "conditioning"; - }; - /** - * ConditioningOutput - * @description Base class for nodes that output a single conditioning tensor - */ - ConditioningOutput: { - /** @description Conditioning tensor */ - conditioning: components["schemas"]["ConditioningField"]; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + /** + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 + */ + denoising_start?: number; + /** + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 + */ + denoising_end?: number; + /** + * Transformer + * @description CogView4 model (Transformer) to load + * @default null + */ + transformer?: components["schemas"]["TransformerField"] | null; + /** + * @description Positive conditioning tensor + * @default null + */ + positive_conditioning?: components["schemas"]["CogView4ConditioningField"] | null; + /** + * @description Negative conditioning tensor + * @default null + */ + negative_conditioning?: components["schemas"]["CogView4ConditioningField"] | null; + /** + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 3.5 + */ + cfg_scale?: number | number[]; + /** + * Width + * @description Width of the generated image. + * @default 1024 + */ + width?: number; + /** + * Height + * @description Height of the generated image. + * @default 1024 + */ + height?: number; + /** + * Steps + * @description Number of steps to run + * @default 25 + */ + steps?: number; + /** + * Seed + * @description Randomness seed for reproducibility. + * @default 0 + */ + seed?: number; /** * type - * @default conditioning_output + * @default cogview4_denoise * @constant */ - type: "conditioning_output"; + type: "cogview4_denoise"; }; /** - * Content Shuffle - * @description Shuffles the image, similar to a 'liquify' filter. + * Image to Latents - CogView4 + * @description Generates latents from an image. */ - ContentShuffleInvocation: { + CogView4ImageToLatentsInvocation: { /** * @description The board to save the image to * @default null @@ -5467,158 +6267,134 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The image to encode. * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Scale Factor - * @description The scale factor used for the shuffle - * @default 256 + * @description VAE + * @default null */ - scale_factor?: number; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default content_shuffle + * @default cogview4_i2l * @constant */ - type: "content_shuffle"; - }; - /** ControlAdapterDefaultSettings */ - ControlAdapterDefaultSettings: { - /** Preprocessor */ - preprocessor: string | null; + type: "cogview4_i2l"; }; - /** ControlField */ - ControlField: { - /** @description The control image */ - image: components["schemas"]["ImageField"]; - /** @description The ControlNet model to use */ - control_model: components["schemas"]["ModelIdentifierField"]; + /** + * Latents to Image - CogView4 + * @description Generates an image from latents. + */ + CogView4LatentsToImageInvocation: { /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * @description The board to save the image to + * @default null */ - control_weight?: number | number[]; + board?: components["schemas"]["BoardField"] | null; /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 + * @description Optional metadata to be saved with the image + * @default null */ - begin_step_percent?: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - end_step_percent?: number; + id: string; /** - * Control Mode - * @description The control mode to use - * @default balanced - * @enum {string} + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + is_intermediate?: boolean; /** - * Resize Mode - * @description The resize mode to use - * @default just_resize - * @enum {string} + * Use Cache + * @description Whether or not to use the cache + * @default true */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; - }; - /** ControlLoRAField */ - ControlLoRAField: { - /** @description Info to load lora model */ - lora: components["schemas"]["ModelIdentifierField"]; + use_cache?: boolean; /** - * Weight - * @description Weight to apply to lora model + * @description Latents tensor + * @default null */ - weight: number; - /** @description Image to use in structural conditioning */ - img: components["schemas"]["ImageField"]; - }; - /** - * ControlLoRA_LyCORIS_FLUX_Config - * @description Model config for Control LoRA models. - */ - ControlLoRA_LyCORIS_FLUX_Config: { + latents?: components["schemas"]["LatentsField"] | null; /** - * Key - * @description A unique key for this model. + * @description VAE + * @default null */ - key: string; + vae?: components["schemas"]["VAEField"] | null; /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. + * type + * @default cogview4_l2i + * @constant */ - file_size: number; + type: "cogview4_l2i"; + }; + /** + * Main Model - CogView4 + * @description Loads a CogView4 base model, outputting its submodels. + */ + CogView4ModelLoaderInvocation: { /** - * Name - * @description Name of the model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - name: string; + id: string; /** - * Description - * @description Model description + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - description: string | null; + is_intermediate?: boolean; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + use_cache?: boolean; + /** @description CogView4 model (Transformer) to load */ + model: components["schemas"]["ModelIdentifierField"]; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * type + * @default cogview4_model_loader + * @constant */ - source_api_response: string | null; + type: "cogview4_model_loader"; + }; + /** + * CogView4ModelLoaderOutput + * @description CogView4 base model loader output. + */ + CogView4ModelLoaderOutput: { /** - * Cover Image - * @description Url for image to preview model + * Transformer + * @description Transformer */ - cover_image: string | null; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + transformer: components["schemas"]["TransformerField"]; /** - * Base - * @default flux - * @constant + * GLM Encoder + * @description GLM (THUDM) tokenizer and text encoder */ - base: "flux"; + glm_encoder: components["schemas"]["GlmEncoderField"]; /** - * Type - * @default control_lora - * @constant + * VAE + * @description VAE */ - type: "control_lora"; + vae: components["schemas"]["VAEField"]; /** - * Format - * @default lycoris + * type + * @default cogview4_model_loader_output * @constant */ - format: "lycoris"; - /** Trigger Phrases */ - trigger_phrases: string[] | null; + type: "cogview4_model_loader_output"; }; /** - * ControlNet - SD1.5, SD2, SDXL - * @description Collects ControlNet info to pass to other nodes + * Prompt - CogView4 + * @description Encodes and preps a prompt for a cogview4 image. */ - ControlNetInvocation: { + CogView4TextEncoderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5637,818 +6413,957 @@ export type components = { */ use_cache?: boolean; /** - * @description The control image + * Prompt + * @description Text prompt to encode. * @default null */ - image?: components["schemas"]["ImageField"] | null; + prompt?: string | null; /** - * @description ControlNet model to load + * GLM Encoder + * @description GLM (THUDM) tokenizer and text encoder * @default null */ - control_model?: components["schemas"]["ModelIdentifierField"] | null; + glm_encoder?: components["schemas"]["GlmEncoderField"] | null; /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * type + * @default cogview4_text_encoder + * @constant */ - control_weight?: number | number[]; + type: "cogview4_text_encoder"; + }; + /** + * CollectInvocation + * @description Collects values into a collection + */ + CollectInvocation: { /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - begin_step_percent?: number; + id: string; /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - end_step_percent?: number; + is_intermediate?: boolean; /** - * Control Mode - * @description The control mode used - * @default balanced - * @enum {string} + * Use Cache + * @description Whether or not to use the cache + * @default true */ - control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + use_cache?: boolean; /** - * Resize Mode - * @description The resize mode used - * @default just_resize - * @enum {string} + * Collection Item + * @description The item to collect (all inputs must be of the same type) + * @default null */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + item?: unknown | null; + /** + * Collection + * @description An optional collection to append to + * @default [] + */ + collection?: unknown[]; /** * type - * @default controlnet + * @default collect * @constant */ - type: "controlnet"; + type: "collect"; }; - /** ControlNetMetadataField */ - ControlNetMetadataField: { - /** @description The control image */ - image: components["schemas"]["ImageField"]; + /** CollectInvocationOutput */ + CollectInvocationOutput: { /** - * @description The control image, after processing. - * @default null + * Collection + * @description The collection of input items */ - processed_image?: components["schemas"]["ImageField"] | null; - /** @description The ControlNet model to use */ - control_model: components["schemas"]["ModelIdentifierField"]; + collection: unknown[]; /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * type + * @default collect_output + * @constant */ - control_weight?: number | number[]; + type: "collect_output"; + }; + /** + * Collection Count + * @description Counts the number of items in a collection. + */ + CollectionCountInvocation: { /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - begin_step_percent?: number; + id: string; /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - end_step_percent?: number; + is_intermediate?: boolean; /** - * Control Mode - * @description The control mode to use - * @default balanced - * @enum {string} + * Use Cache + * @description Whether or not to use the cache + * @default false */ - control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + use_cache?: boolean; /** - * Resize Mode - * @description The resize mode to use - * @default just_resize - * @enum {string} + * Collection + * @description The collection to count + * @default [] */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + collection?: unknown[]; + /** + * type + * @default collection_count + * @constant + */ + type: "collection_count"; }; /** - * ControlNetRecallParameter - * @description ControlNet configuration for recall + * CollectionCountOutput + * @description The output of the collection count node. */ - ControlNetRecallParameter: { + CollectionCountOutput: { /** - * Model Name - * @description The name of the ControlNet/T2I Adapter/Control LoRA model + * Count + * @description The number of items in the collection */ - model_name: string; + count: number; /** - * Image Name - * @description The filename of the control image in outputs/images + * type + * @default collection_count_output + * @constant */ - image_name?: string | null; + type: "collection_count_output"; + }; + /** + * Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option + */ + CollectionIndexInvocation: { /** - * Weight - * @description The weight for the control adapter - * @default 1 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - weight?: number; + id: string; /** - * Begin Step Percent - * @description When the control adapter is first applied (% of total steps) + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - begin_step_percent?: number | null; + is_intermediate?: boolean; /** - * End Step Percent - * @description When the control adapter is last applied (% of total steps) + * Use Cache + * @description Whether or not to use the cache + * @default false */ - end_step_percent?: number | null; + use_cache?: boolean; /** - * Control Mode - * @description The control mode (ControlNet only) + * Random + * @description Random Index? + * @default true */ - control_mode?: ("balanced" | "more_prompt" | "more_control") | null; - }; - /** ControlNet_Checkpoint_FLUX_Config */ - ControlNet_Checkpoint_FLUX_Config: { + random?: boolean; /** - * Key - * @description A unique key for this model. + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 */ - key: string; + index?: number; /** - * Hash - * @description The hash of the model file(s). + * Collection + * @description collection + * @default null */ - hash: string; + collection?: unknown[] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * type + * @default collection_index + * @constant */ - path: string; + type: "collection_index"; + }; + /** + * CollectionIndexOutput + * @description Used to connect iteration outputs. Will be expanded to a specific output. + */ + CollectionIndexOutput: { /** - * File Size - * @description The size of the model in bytes. + * Collection Item + * @description The item being iterated over */ - file_size: number; + item: unknown; /** - * Name - * @description Name of the model. + * Index + * @description The index of the selected item */ - name: string; + index: number; /** - * Description - * @description Model description + * Total + * @description The total number of items in the collection */ - description: string | null; + total: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default collection_index_output + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "collection_index_output"; + }; + /** + * Collection Join + * @description CollectionJoin Joins two collections into a single collection + */ + CollectionJoinInvocation: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source_api_response: string | null; + id: string; /** - * Cover Image - * @description Url for image to preview model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cover_image: string | null; + is_intermediate?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * Use Cache + * @description Whether or not to use the cache + * @default false */ - config_path: string | null; + use_cache?: boolean; /** - * Type - * @default controlnet - * @constant + * Collection A + * @description collection + * @default [] */ - type: "controlnet"; + collection_a?: unknown[]; /** - * Format - * @default checkpoint - * @constant + * Collection B + * @description collection + * @default [] */ - format: "checkpoint"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + collection_b?: unknown[]; /** - * Base - * @default flux + * type + * @default collection_join * @constant */ - base: "flux"; + type: "collection_join"; }; - /** ControlNet_Checkpoint_SD1_Config */ - ControlNet_Checkpoint_SD1_Config: { + /** CollectionJoinOutput */ + CollectionJoinOutput: { /** - * Key - * @description A unique key for this model. + * Collection + * @description The collection of output items */ - key: string; + collection: unknown[]; /** - * Hash - * @description The hash of the model file(s). + * type + * @default collection_join_output + * @constant */ - hash: string; + type: "collection_join_output"; + }; + /** + * Collection Reverse + * @description Reverses a collection. + */ + CollectionReverseInvocation: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default false */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * Collection + * @description The collection to reverse + * @default [] */ - description: string | null; + collection?: unknown[]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default collection_reverse + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "collection_reverse"; + }; + /** + * CollectionReverseOutput + * @description The output of the collection reverse node. + */ + CollectionReverseOutput: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Collection + * @description The reversed collection */ - source_api_response: string | null; + collection: unknown[]; /** - * Cover Image - * @description Url for image to preview model + * type + * @default collection_reverse_output + * @constant */ - cover_image: string | null; + type: "collection_reverse_output"; + }; + /** + * Collection Slice + * @description Slices a collection. + */ + CollectionSliceInvocation: { /** - * Config Path - * @description Path to the config for this model, if any. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - config_path: string | null; + id: string; /** - * Type - * @default controlnet - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - type: "controlnet"; + is_intermediate?: boolean; /** - * Format - * @default checkpoint - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default false */ - format: "checkpoint"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + use_cache?: boolean; /** - * Base - * @default sd-1 - * @constant + * Collection + * @description The collection to slice + * @default [] */ - base: "sd-1"; - }; - /** ControlNet_Checkpoint_SD2_Config */ - ControlNet_Checkpoint_SD2_Config: { + collection?: unknown[]; /** - * Key - * @description A unique key for this model. + * Start + * @description The start index of the slice + * @default 0 */ - key: string; + start?: number; /** - * Hash - * @description The hash of the model file(s). + * Stop + * @description The stop index of the slice (exclusive) + * @default null */ - hash: string; + stop?: number | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Step + * @description The step of the slice + * @default 1 */ - path: string; + step?: number; /** - * File Size - * @description The size of the model in bytes. + * type + * @default collection_slice + * @constant */ - file_size: number; + type: "collection_slice"; + }; + /** + * CollectionSliceOutput + * @description The output of the collection slice node. + */ + CollectionSliceOutput: { /** - * Name - * @description Name of the model. + * Collection + * @description The sliced collection */ - name: string; + collection: unknown[]; /** - * Description - * @description Model description + * type + * @default collection_slice_output + * @constant */ - description: string | null; + type: "collection_slice_output"; + }; + /** + * Collection Sort + * @description CollectionSort Sorts a collection + */ + CollectionSortInvocation: { /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default false */ - cover_image: string | null; + use_cache?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * Collection + * @description collection + * @default [] */ - config_path: string | null; + collection?: unknown[]; /** - * Type - * @default controlnet - * @constant + * Reverse + * @description Reverse Sort + * @default false */ - type: "controlnet"; + reverse?: boolean; /** - * Format - * @default checkpoint + * type + * @default collection_sort * @constant */ - format: "checkpoint"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + type: "collection_sort"; + }; + /** CollectionSortOutput */ + CollectionSortOutput: { /** - * Base - * @default sd-2 + * Collection + * @description The collection of output items + */ + collection: unknown[]; + /** + * type + * @default collection_sort_output * @constant */ - base: "sd-2"; + type: "collection_sort_output"; }; - /** ControlNet_Checkpoint_SDXL_Config */ - ControlNet_Checkpoint_SDXL_Config: { + /** + * Collection Unique + * @description Removes duplicate items from a collection. + */ + CollectionUniqueInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default false */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Collection + * @description The collection to deduplicate + * @default [] */ - file_size: number; + collection?: unknown[]; /** - * Name - * @description Name of the model. + * type + * @default collection_unique + * @constant */ - name: string; + type: "collection_unique"; + }; + /** + * CollectionUniqueOutput + * @description The output of the collection unique node. + */ + CollectionUniqueOutput: { /** - * Description - * @description Model description + * Collection + * @description The collection with unique items */ - description: string | null; + collection: unknown[]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default collection_unique_output + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "collection_unique_output"; + }; + /** + * Color Cast Correction + * @description Correct color cast while preserving perceptual brightness + */ + ColorCastCorrectionInvocation: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description Optional metadata to be saved with the image + * @default null */ - source_api_response: string | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Cover Image - * @description Url for image to preview model + * @description The board to save the image to + * @default null */ - cover_image: string | null; + board?: components["schemas"]["BoardField"] | null; /** - * Config Path - * @description Path to the config for this model, if any. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - config_path: string | null; + id: string; /** - * Type - * @default controlnet - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - type: "controlnet"; + is_intermediate?: boolean; /** - * Format - * @default checkpoint - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - format: "checkpoint"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + use_cache?: boolean; /** - * Base - * @default sdxl + * @description Image to correct + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Strength + * @description Strength of correction adjustment + * @default 0.75 + */ + strength?: number; + /** + * @description The target color for correction + * @default { + * "r": 127, + * "g": 127, + * "b": 127, + * "a": 255 + * } + */ + target_color?: components["schemas"]["ColorField"]; + /** + * type + * @default color_cast_correction * @constant */ - base: "sdxl"; + type: "color_cast_correction"; }; /** - * ControlNet_Checkpoint_ZImage_Config - * @description Model config for Z-Image Control adapter models (Safetensors checkpoint). - * - * Z-Image Control models are standalone adapters containing only the control layers - * (control_layers, control_all_x_embedder, control_noise_refiner) that extend - * the base Z-Image transformer with spatial conditioning capabilities. - * - * Supports: Canny, HED, Depth, Pose, MLSD. - * Recommended control_context_scale: 0.65-0.80. + * ColorCollectionOutput + * @description Base class for nodes that output a collection of colors */ - ControlNet_Checkpoint_ZImage_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; + ColorCollectionOutput: { /** - * Hash - * @description The hash of the model file(s). + * Collection + * @description The output colors */ - hash: string; + collection: components["schemas"]["ColorField"][]; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * type + * @default color_collection_output + * @constant */ - path: string; + type: "color_collection_output"; + }; + /** + * Color Correct + * @description Matches the color histogram of a base image to a reference image, optionally + * using a mask to only color-correct certain regions of the base image. + */ + ColorCorrectInvocation: { /** - * File Size - * @description The size of the model in bytes. + * @description The board to save the image to + * @default null */ - file_size: number; + board?: components["schemas"]["BoardField"] | null; /** - * Name - * @description Name of the model. + * @description Optional metadata to be saved with the image + * @default null */ - name: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * @description The image to color-correct + * @default null */ - cover_image: string | null; + base_image?: components["schemas"]["ImageField"] | null; /** - * Config Path - * @description Path to the config for this model, if any. + * @description Reference image for color-correction + * @default null */ - config_path: string | null; + color_reference?: components["schemas"]["ImageField"] | null; /** - * Type - * @default controlnet - * @constant + * @description Optional mask to limit color correction area + * @default null */ - type: "controlnet"; + mask?: components["schemas"]["ImageField"] | null; /** - * Format - * @default checkpoint - * @constant + * Color Space + * @description Colorspace in which to apply histogram matching + * @default RGB + * @enum {string} */ - format: "checkpoint"; + colorspace?: "RGB" | "YCbCr" | "YCbCr-Chroma" | "YCbCr-Luma"; /** - * Base - * @default z-image + * type + * @default color_correct * @constant */ - base: "z-image"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + type: "color_correct"; }; - /** ControlNet_Diffusers_FLUX_Config */ - ControlNet_Diffusers_FLUX_Config: { + /** + * ColorField + * @description A color primitive field + */ + ColorField: { /** - * Key - * @description A unique key for this model. + * R + * @description The red component */ - key: string; + r: number; /** - * Hash - * @description The hash of the model file(s). + * G + * @description The green component */ - hash: string; + g: number; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * B + * @description The blue component */ - path: string; + b: number; /** - * File Size - * @description The size of the model in bytes. + * A + * @description The alpha component */ - file_size: number; + a: number; + }; + /** + * Color Primitive + * @description A color primitive value + */ + ColorInvocation: { /** - * Name - * @description Name of the model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - name: string; + id: string; /** - * Description - * @description Model description + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - description: string | null; + is_intermediate?: boolean; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + use_cache?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description The color value + * @default { + * "r": 0, + * "g": 0, + * "b": 0, + * "a": 255 + * } */ - source_api_response: string | null; + color?: components["schemas"]["ColorField"]; /** - * Cover Image - * @description Url for image to preview model + * type + * @default color + * @constant */ - cover_image: string | null; + type: "color"; + }; + /** + * Color Map + * @description Generates a color map from the provided image. + */ + ColorMapInvocation: { /** - * Format - * @default diffusers - * @constant + * @description The board to save the image to + * @default null */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + board?: components["schemas"]["BoardField"] | null; /** - * Type - * @default controlnet - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "controlnet"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Base - * @default flux - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - base: "flux"; - }; - /** ControlNet_Diffusers_SD1_Config */ - ControlNet_Diffusers_SD1_Config: { + id: string; /** - * Key - * @description A unique key for this model. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - key: string; + is_intermediate?: boolean; /** - * Hash - * @description The hash of the model file(s). + * Use Cache + * @description Whether or not to use the cache + * @default true */ - hash: string; + use_cache?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * @description The image to process + * @default null */ - path: string; + image?: components["schemas"]["ImageField"] | null; /** - * File Size - * @description The size of the model in bytes. + * Tile Size + * @description Tile size + * @default 64 */ - file_size: number; + tile_size?: number; /** - * Name - * @description Name of the model. + * type + * @default color_map + * @constant */ - name: string; + type: "color_map"; + }; + /** + * ColorOutput + * @description Base class for nodes that output a single color + */ + ColorOutput: { + /** @description The output color */ + color: components["schemas"]["ColorField"]; /** - * Description - * @description Model description + * type + * @default color_output + * @constant */ - description: string | null; + type: "color_output"; + }; + /** + * Compare Floats + * @description Compares two floats based on input criteria and ouputs a boolean. + */ + CompareFloatsInvocation: { /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Format - * @default diffusers - * @constant + * Comparison Method + * @description The comparision method to use + * @default null */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + comparison_method?: ("<" | "<=" | ">" | ">=" | "==" | "!=") | null; /** - * Type - * @default controlnet - * @constant + * Float 1 + * @description The first float + * @default null */ - type: "controlnet"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + float1?: number | null; /** - * Base - * @default sd-1 + * Float 2 + * @description The second float + * @default null + */ + float2?: number | null; + /** + * type + * @default compare_floats_invocation * @constant */ - base: "sd-1"; + type: "compare_floats_invocation"; }; - /** ControlNet_Diffusers_SD2_Config */ - ControlNet_Diffusers_SD2_Config: { + /** + * Compare Ints + * @description Compares two integers based on input criteria and ouputs a boolean. + */ + CompareIntsInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Comparison Method + * @description The comparision method to use + * @default null */ - file_size: number; + comparison_method?: ("<" | "<=" | ">" | ">=" | "==" | "!=") | null; /** - * Name - * @description Name of the model. + * Int 1 + * @description The first integer. + * @default null */ - name: string; + int1?: number | null; /** - * Description - * @description Model description + * Int 2 + * @description The second integer. + * @default null */ - description: string | null; + int2?: number | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default compare_ints_invocation + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "compare_ints_invocation"; + }; + /** + * Compare Strings + * @description Compares two strings based on input criteria and ouputs a boolean. + */ + CompareStringsInvocation: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source_api_response: string | null; + id: string; /** - * Cover Image - * @description Url for image to preview model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cover_image: string | null; + is_intermediate?: boolean; /** - * Format - * @default diffusers - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + use_cache?: boolean; /** - * Type - * @default controlnet - * @constant + * Comparison Method + * @description The comparision method to use + * @default null */ - type: "controlnet"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + comparison_method?: ("equals" | "contains" | "starts with" | "ends with") | null; /** - * Base - * @default sd-2 - * @constant - */ - base: "sd-2"; - }; - /** ControlNet_Diffusers_SDXL_Config */ - ControlNet_Diffusers_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Ignore Case + * @description If true the node will ignore the case of the strings in all comparison methods + * @default null */ - path: string; + ignore_case?: boolean | null; /** - * File Size - * @description The size of the model in bytes. + * String 1 + * @description The first float + * @default null */ - file_size: number; + str1?: string | null; /** - * Name - * @description Name of the model. + * String 2 + * @description The second float + * @default null */ - name: string; + str2?: string | null; /** - * Description - * @description Model description + * type + * @default compare_strings_invocation + * @constant */ - description: string | null; + type: "compare_strings_invocation"; + }; + /** + * Prompt - SD1.5 + * @description Parse prompt using compel package to conditioning. + */ + CompelInvocation: { /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Format - * @default diffusers - * @constant + * Prompt + * @description Prompt to be parsed by Compel to create a conditioning tensor + * @default */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + prompt?: string; /** - * Type - * @default controlnet - * @constant + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - type: "controlnet"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + clip?: components["schemas"]["CLIPField"] | null; /** - * Base - * @default sdxl - * @constant + * @description A mask defining the region that this conditioning prompt applies to. + * @default null */ - base: "sdxl"; - }; - /** - * ControlOutput - * @description node output for ControlNet info - */ - ControlOutput: { - /** @description ControlNet(s) to apply */ - control: components["schemas"]["ControlField"]; + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default control_output + * @default compel * @constant */ - type: "control_output"; + type: "compel"; }; /** - * Core Metadata - * @description Used internally by Invoke to collect metadata for generations. + * Concatenate Flux Conditionings + * @description Concatenates the T5 embedding tensors of up to six input Flux Conditioning objects. + * Provides flexible control over the CLIP embedding: select by 1-indexed input number, + * or generate a zeros tensor. */ - CoreMetadataInvocation: { + ConcatenateFluxConditioningInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -6467,281 +7382,181 @@ export type components = { */ use_cache?: boolean; /** - * Generation Mode - * @description The generation mode that output this image - * @default null - */ - generation_mode?: ("txt2img" | "img2img" | "inpaint" | "outpaint" | "sdxl_txt2img" | "sdxl_img2img" | "sdxl_inpaint" | "sdxl_outpaint" | "flux_txt2img" | "flux_img2img" | "flux_inpaint" | "flux_outpaint" | "flux2_txt2img" | "flux2_img2img" | "flux2_inpaint" | "flux2_outpaint" | "sd3_txt2img" | "sd3_img2img" | "sd3_inpaint" | "sd3_outpaint" | "cogview4_txt2img" | "cogview4_img2img" | "cogview4_inpaint" | "cogview4_outpaint" | "z_image_txt2img" | "z_image_img2img" | "z_image_inpaint" | "z_image_outpaint") | null; - /** - * Positive Prompt - * @description The positive prompt parameter + * @description First optional Flux Conditioning input. * @default null */ - positive_prompt?: string | null; + conditioning_1?: components["schemas"]["FluxConditioningField"] | null; /** - * Negative Prompt - * @description The negative prompt parameter - * @default null + * Strength 1 + * @description Strength for the first conditioning input (multiplies its embedding tensors). + * @default 1 */ - negative_prompt?: string | null; + strength_1?: number; /** - * Width - * @description The width parameter + * @description Second optional Flux Conditioning input. * @default null */ - width?: number | null; + conditioning_2?: components["schemas"]["FluxConditioningField"] | null; /** - * Height - * @description The height parameter - * @default null + * Strength 2 + * @description Strength for the second conditioning input (multiplies its embedding tensors). + * @default 1 */ - height?: number | null; + strength_2?: number; /** - * Seed - * @description The seed used for noise generation + * @description Third optional Flux Conditioning input. * @default null */ - seed?: number | null; + conditioning_3?: components["schemas"]["FluxConditioningField"] | null; /** - * Rand Device - * @description The device used for random number generation - * @default null + * Strength 3 + * @description Strength for the third conditioning input (multiplies its embedding tensors). + * @default 1 */ - rand_device?: string | null; + strength_3?: number; /** - * Cfg Scale - * @description The classifier-free guidance scale parameter + * @description Fourth optional Flux Conditioning input. * @default null */ - cfg_scale?: number | null; + conditioning_4?: components["schemas"]["FluxConditioningField"] | null; /** - * Cfg Rescale Multiplier - * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR - * @default null + * Strength 4 + * @description Strength for the fourth conditioning input (multiplies its embedding tensors). + * @default 1 */ - cfg_rescale_multiplier?: number | null; + strength_4?: number; /** - * Steps - * @description The number of steps used for inference + * @description Fifth optional Flux Conditioning input. * @default null */ - steps?: number | null; + conditioning_5?: components["schemas"]["FluxConditioningField"] | null; /** - * Scheduler - * @description The scheduler used for inference - * @default null + * Strength 5 + * @description Strength for the fifth conditioning input (multiplies its embedding tensors). + * @default 1 */ - scheduler?: string | null; + strength_5?: number; /** - * Seamless X - * @description Whether seamless tiling was used on the X axis + * @description Sixth optional Flux Conditioning input. * @default null */ - seamless_x?: boolean | null; + conditioning_6?: components["schemas"]["FluxConditioningField"] | null; /** - * Seamless Y - * @description Whether seamless tiling was used on the Y axis - * @default null + * Strength 6 + * @description Strength for the sixth conditioning input (multiplies its embedding tensors). + * @default 1 */ - seamless_y?: boolean | null; + strength_6?: number; /** - * Clip Skip - * @description The number of skipped CLIP layers - * @default null + * Select Clip + * @description CLIP embedding selection: 0 for a zeros tensor; 1-6 to select a specific input (1-indexed). If a selected input is missing, it falls back to the next subsequent, then preceding, available CLIP embedding. + * @default 1 */ - clip_skip?: number | null; + select_clip?: number; /** - * @description The main model used for inference - * @default null + * type + * @default flux_conditioning_concatenate + * @constant */ - model?: components["schemas"]["ModelIdentifierField"] | null; + type: "flux_conditioning_concatenate"; + }; + /** + * Conditioning Collection Primitive + * @description A collection of conditioning tensor primitive values + */ + ConditioningCollectionInvocation: { /** - * Controlnets - * @description The ControlNets used for inference - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - controlnets?: components["schemas"]["ControlNetMetadataField"][] | null; + id: string; /** - * Ipadapters - * @description The IP Adapters used for inference - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - ipAdapters?: components["schemas"]["IPAdapterMetadataField"][] | null; + is_intermediate?: boolean; /** - * T2Iadapters - * @description The IP Adapters used for inference - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - t2iAdapters?: components["schemas"]["T2IAdapterMetadataField"][] | null; + use_cache?: boolean; /** - * Loras - * @description The LoRAs used for inference - * @default null + * Collection + * @description The collection of conditioning tensors + * @default [] */ - loras?: components["schemas"]["LoRAMetadataField"][] | null; + collection?: components["schemas"]["ConditioningField"][]; /** - * Strength - * @description The strength used for latents-to-latents - * @default null + * type + * @default conditioning_collection + * @constant */ - strength?: number | null; + type: "conditioning_collection"; + }; + /** + * Conditioning Collection Primitive Linked + * @description A collection of conditioning tensor primitive values + */ + ConditioningCollectionLinkedInvocation: { /** - * Init Image - * @description The name of the initial image - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - init_image?: string | null; + id: string; /** - * @description The VAE used for decoding, if the main model's default was not used - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - vae?: components["schemas"]["ModelIdentifierField"] | null; + is_intermediate?: boolean; /** - * @description The Qwen3 text encoder model used for Z-Image inference - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - qwen3_encoder?: components["schemas"]["ModelIdentifierField"] | null; + use_cache?: boolean; /** - * Hrf Enabled - * @description Whether or not high resolution fix was enabled. - * @default null + * Collection + * @description The collection of conditioning tensors + * @default [] */ - hrf_enabled?: boolean | null; + collection?: components["schemas"]["ConditioningField"][]; /** - * Hrf Method - * @description The high resolution fix upscale method. - * @default null + * type + * @default conditioning_collection_linked + * @constant */ - hrf_method?: string | null; + type: "conditioning_collection_linked"; /** - * Hrf Strength - * @description The high resolution fix img2img strength used in the upscale pass. + * @description Conditioning tensor * @default null */ - hrf_strength?: number | null; - /** - * Positive Style Prompt - * @description The positive style prompt parameter - * @default null - */ - positive_style_prompt?: string | null; - /** - * Negative Style Prompt - * @description The negative style prompt parameter - * @default null - */ - negative_style_prompt?: string | null; - /** - * @description The SDXL Refiner model used - * @default null - */ - refiner_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Refiner Cfg Scale - * @description The classifier-free guidance scale parameter used for the refiner - * @default null - */ - refiner_cfg_scale?: number | null; - /** - * Refiner Steps - * @description The number of steps used for the refiner - * @default null - */ - refiner_steps?: number | null; - /** - * Refiner Scheduler - * @description The scheduler used for the refiner - * @default null - */ - refiner_scheduler?: string | null; - /** - * Refiner Positive Aesthetic Score - * @description The aesthetic score used for the refiner - * @default null - */ - refiner_positive_aesthetic_score?: number | null; - /** - * Refiner Negative Aesthetic Score - * @description The aesthetic score used for the refiner - * @default null - */ - refiner_negative_aesthetic_score?: number | null; - /** - * Refiner Start - * @description The start value used for refiner denoising - * @default null - */ - refiner_start?: number | null; - /** - * type - * @default core_metadata - * @constant - */ - type: "core_metadata"; - } & { - [key: string]: unknown; + conditioning?: components["schemas"]["ConditioningField"] | null; }; /** - * Create Denoise Mask - * @description Creates mask for denoising model run. + * ConditioningCollectionOutput + * @description Base class for nodes that output a collection of conditioning tensors */ - CreateDenoiseMaskInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description VAE - * @default null - */ - vae?: components["schemas"]["VAEField"] | null; - /** - * @description Image which will be masked - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The mask to use when pasting - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Tiled - * @description Processing using overlapping tiles (reduce memory consumption) - * @default false - */ - tiled?: boolean; + ConditioningCollectionOutput: { /** - * Fp32 - * @description Whether or not to use full float32 precision - * @default false + * Collection + * @description The output conditioning tensors */ - fp32?: boolean; + collection: components["schemas"]["ConditioningField"][]; /** * type - * @default create_denoise_mask + * @default conditioning_collection_output * @constant */ - type: "create_denoise_mask"; + type: "conditioning_collection_output"; }; /** - * Create Gradient Mask - * @description Creates mask for denoising. + * Conditioning Collection Toggle + * @description Allows boolean selection between two separate conditioning collection inputs */ - CreateGradientMaskInvocation: { + ConditioningCollectionToggleInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -6760,80 +7575,51 @@ export type components = { */ use_cache?: boolean; /** - * @description Image which will be masked - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Edge Radius - * @description How far to expand the edges of the mask - * @default 16 - */ - edge_radius?: number; - /** - * Coherence Mode - * @default Gaussian Blur - * @enum {string} - */ - coherence_mode?: "Gaussian Blur" | "Box Blur" | "Staged"; - /** - * Minimum Denoise - * @description Minimum denoise level for the coherence region - * @default 0 - */ - minimum_denoise?: number; - /** - * [OPTIONAL] Image - * @description OPTIONAL: Only connect for specialized Inpainting models, masked_latents will be generated from the image with the VAE - * @default null + * Use Second + * @description Use 2nd Input + * @default false */ - image?: components["schemas"]["ImageField"] | null; + use_second?: boolean; /** - * [OPTIONAL] UNet - * @description OPTIONAL: If the Unet is a specialized Inpainting model, masked_latents will be generated from the image with the VAE + * Col1 + * @description First Conditioning Collection Input * @default null */ - unet?: components["schemas"]["UNetField"] | null; + col1?: components["schemas"]["ConditioningField"][] | null; /** - * [OPTIONAL] VAE - * @description OPTIONAL: Only connect for specialized Inpainting models, masked_latents will be generated from the image with the VAE + * Col2 + * @description Second Conditioning Collection Input * @default null */ - vae?: components["schemas"]["VAEField"] | null; - /** - * Tiled - * @description Processing using overlapping tiles (reduce memory consumption) - * @default false - */ - tiled?: boolean; - /** - * Fp32 - * @description Whether or not to use full float32 precision - * @default false - */ - fp32?: boolean; + col2?: components["schemas"]["ConditioningField"][] | null; /** * type - * @default create_gradient_mask + * @default conditioning_collection_toggle * @constant */ - type: "create_gradient_mask"; + type: "conditioning_collection_toggle"; }; /** - * Crop Image to Bounding Box - * @description Crop an image to the given bounding box. If the bounding box is omitted, the image is cropped to the non-transparent pixels. + * ConditioningField + * @description A conditioning tensor primitive value */ - CropImageToBoundingBoxInvocation: { + ConditioningField: { /** - * @description The board to save the image to - * @default null + * Conditioning Name + * @description The name of conditioning tensor */ - board?: components["schemas"]["BoardField"] | null; + conditioning_name: string; /** - * @description Optional metadata to be saved with the image + * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + mask?: components["schemas"]["TensorField"] | null; + }; + /** + * Conditioning Primitive + * @description A conditioning tensor primitive value + */ + ConditioningInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -6852,28 +7638,36 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to crop + * @description Conditioning tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; + conditioning?: components["schemas"]["ConditioningField"] | null; /** - * @description The bounding box to crop the image to - * @default null + * type + * @default conditioning + * @constant */ - bounding_box?: components["schemas"]["BoundingBoxField"] | null; + type: "conditioning"; + }; + /** + * ConditioningOutput + * @description Base class for nodes that output a single conditioning tensor + */ + ConditioningOutput: { + /** @description Conditioning tensor */ + conditioning: components["schemas"]["ConditioningField"]; /** * type - * @default crop_image_to_bounding_box + * @default conditioning_output * @constant */ - type: "crop_image_to_bounding_box"; + type: "conditioning_output"; }; /** - * Crop Latents - * @description Crops a latent-space tensor to a box specified in image-space. The box dimensions and coordinates must be - * divisible by the latent scale factor of 8. + * Conditioning Toggle + * @description Allows boolean selection between two separate conditioning inputs */ - CropLatentsCoreInvocation: { + ConditioningToggleInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -6892,46 +7686,33 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * X - * @description The left x coordinate (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. - * @default null - */ - x?: number | null; - /** - * Y - * @description The top y coordinate (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. - * @default null + * Use Second + * @description Use 2nd Input + * @default false */ - y?: number | null; + use_second?: boolean; /** - * Width - * @description The width (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @description First Conditioning Input * @default null */ - width?: number | null; + cond1?: components["schemas"]["ConditioningField"] | null; /** - * Height - * @description The height (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @description Second Conditioning Input * @default null */ - height?: number | null; + cond2?: components["schemas"]["ConditioningField"] | null; /** * type - * @default crop_latents + * @default conditioning_toggle * @constant */ - type: "crop_latents"; + type: "conditioning_toggle"; }; /** - * OpenCV Inpaint - * @description Simple inpaint using opencv. + * Content Shuffle + * @description Shuffles the image, similar to a 'liquify' filter. */ - CvInpaintInvocation: { + ContentShuffleInvocation: { /** * @description The board to save the image to * @default null @@ -6960,37 +7741,172 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to inpaint + * @description The image to process * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description The mask to use when inpainting - * @default null + * Scale Factor + * @description The scale factor used for the shuffle + * @default 256 */ - mask?: components["schemas"]["ImageField"] | null; + scale_factor?: number; /** * type - * @default cv_inpaint + * @default content_shuffle * @constant */ - type: "cv_inpaint"; + type: "content_shuffle"; + }; + /** ControlAdapterDefaultSettings */ + ControlAdapterDefaultSettings: { + /** Preprocessor */ + preprocessor: string | null; + }; + /** ControlField */ + ControlField: { + /** @description The control image */ + image: components["schemas"]["ImageField"]; + /** @description The ControlNet model to use */ + control_model: components["schemas"]["ModelIdentifierField"]; + /** + * Control Weight + * @description The weight given to the ControlNet + * @default 1 + */ + control_weight?: number | number[]; + /** + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 + */ + begin_step_percent?: number; + /** + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; + /** + * Control Mode + * @description The control mode to use + * @default balanced + * @enum {string} + */ + control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + /** + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} + */ + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + }; + /** ControlListOutput */ + ControlListOutput: { + /** + * ControlNet List + * @description ControlNet(s) to apply + */ + control_list: components["schemas"]["ControlField"][]; + /** + * type + * @default control_list_output + * @constant + */ + type: "control_list_output"; + }; + /** ControlLoRAField */ + ControlLoRAField: { + /** @description Info to load lora model */ + lora: components["schemas"]["ModelIdentifierField"]; + /** + * Weight + * @description Weight to apply to lora model + */ + weight: number; + /** @description Image to use in structural conditioning */ + img: components["schemas"]["ImageField"]; }; /** - * DW Openpose Detection - * @description Generates an openpose pose from an image using DWPose + * ControlLoRA_LyCORIS_FLUX_Config + * @description Model config for Control LoRA models. */ - DWOpenposeDetectionInvocation: { + ControlLoRA_LyCORIS_FLUX_Config: { /** - * @description The board to save the image to - * @default null + * Key + * @description A unique key for this model. */ - board?: components["schemas"]["BoardField"] | null; + key: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Hash + * @description The hash of the model file(s). */ - metadata?: components["schemas"]["MetadataField"] | null; + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; + /** + * Type + * @default control_lora + * @constant + */ + type: "control_lora"; + /** + * Format + * @default lycoris + * @constant + */ + format: "lycoris"; + /** Trigger Phrases */ + trigger_phrases: string[] | null; + }; + /** + * ControlNet - SD1.5, SD2, SDXL + * @description Collects ControlNet info to pass to other nodes + */ + ControlNetInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -7009,37 +7925,59 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The control image * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Draw Body - * @default true + * @description ControlNet model to load + * @default null */ - draw_body?: boolean; + control_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Draw Face - * @default false + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - draw_face?: boolean; + control_weight?: number | number[]; /** - * Draw Hands - * @default false + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - draw_hands?: boolean; + begin_step_percent?: number; + /** + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; + /** + * Control Mode + * @description The control mode used + * @default balanced + * @enum {string} + */ + control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + /** + * Resize Mode + * @description The resize mode used + * @default just_resize + * @enum {string} + */ + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** * type - * @default dw_openpose_detection + * @default controlnet * @constant */ - type: "dw_openpose_detection"; + type: "controlnet"; }; /** - * Decode Invisible Watermark - * @description Decode an invisible watermark from an image. + * ControlNet-Linked + * @description Collects ControlNet info to pass to other nodes. */ - DecodeInvisibleWatermarkInvocation: { + ControlNetLinkedInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -7058,913 +7996,831 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to decode the watermark from + * @description The control image * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Length - * @description The expected watermark length in bytes - * @default 8 + * @description ControlNet model to load + * @default null */ - length?: number; + control_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * type - * @default decode_watermark - * @constant + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - type: "decode_watermark"; - }; - /** - * DeleteAllExceptCurrentResult - * @description Result of deleting all except current - */ - DeleteAllExceptCurrentResult: { + control_weight?: number | number[]; /** - * Deleted - * @description Number of queue items deleted + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - deleted: number; - }; - /** DeleteBoardResult */ - DeleteBoardResult: { + begin_step_percent?: number; /** - * Board Id - * @description The id of the board that was deleted. + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - board_id: string; + end_step_percent?: number; /** - * Deleted Board Images - * @description The image names of the board-images relationships that were deleted. + * Control Mode + * @description The control mode used + * @default balanced + * @enum {string} */ - deleted_board_images: string[]; + control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; /** - * Deleted Images - * @description The names of the images that were deleted. + * Resize Mode + * @description The resize mode used + * @default just_resize + * @enum {string} */ - deleted_images: string[]; - }; - /** - * DeleteByDestinationResult - * @description Result of deleting by a destination - */ - DeleteByDestinationResult: { + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** - * Deleted - * @description Number of queue items deleted + * type + * @default controlnet-linked + * @constant */ - deleted: number; + type: "controlnet-linked"; + /** + * ControlNet-List + * @description ControlNet(s) to apply + * @default null + */ + control_list?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; }; - /** DeleteImagesResult */ - DeleteImagesResult: { + /** ControlNetMetadataField */ + ControlNetMetadataField: { + /** @description The control image */ + image: components["schemas"]["ImageField"]; /** - * Affected Boards - * @description The ids of boards affected by the delete operation + * @description The control image, after processing. + * @default null */ - affected_boards: string[]; + processed_image?: components["schemas"]["ImageField"] | null; + /** @description The ControlNet model to use */ + control_model: components["schemas"]["ModelIdentifierField"]; /** - * Deleted Images - * @description The names of the images that were deleted + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - deleted_images: string[]; - }; - /** - * DeleteOrphanedModelsRequest - * @description Request to delete specific orphaned model directories. - */ - DeleteOrphanedModelsRequest: { + control_weight?: number | number[]; /** - * Paths - * @description List of relative paths to delete + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - paths: string[]; - }; - /** - * DeleteOrphanedModelsResponse - * @description Response from deleting orphaned models. - */ - DeleteOrphanedModelsResponse: { + begin_step_percent?: number; /** - * Deleted - * @description Paths that were successfully deleted + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - deleted: string[]; + end_step_percent?: number; /** - * Errors - * @description Paths that had errors, with error messages + * Control Mode + * @description The control mode to use + * @default balanced + * @enum {string} */ - errors: { - [key: string]: string; - }; + control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + /** + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} + */ + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; }; /** - * Denoise - SD1.5, SDXL - * @description Denoises noisy latents to decodable images + * ControlNetRecallParameter + * @description ControlNet configuration for recall */ - DenoiseLatentsInvocation: { + ControlNetRecallParameter: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Model Name + * @description The name of the ControlNet/T2I Adapter/Control LoRA model */ - id: string; + model_name: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Image Name + * @description The filename of the control image in outputs/images */ - is_intermediate?: boolean; + image_name?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Weight + * @description The weight for the control adapter + * @default 1 */ - use_cache?: boolean; + weight?: number; /** - * Positive Conditioning - * @description Positive conditioning tensor - * @default null + * Begin Step Percent + * @description When the control adapter is first applied (% of total steps) */ - positive_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; + begin_step_percent?: number | null; /** - * Negative Conditioning - * @description Negative conditioning tensor - * @default null + * End Step Percent + * @description When the control adapter is last applied (% of total steps) */ - negative_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; + end_step_percent?: number | null; /** - * @description Noise tensor - * @default null + * Control Mode + * @description The control mode (ControlNet only) */ - noise?: components["schemas"]["LatentsField"] | null; + control_mode?: ("balanced" | "more_prompt" | "more_control") | null; + }; + /** ControlNet_Checkpoint_FLUX_Config */ + ControlNet_Checkpoint_FLUX_Config: { /** - * Steps - * @description Number of steps to run - * @default 10 + * Key + * @description A unique key for this model. */ - steps?: number; + key: string; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 7.5 + * Hash + * @description The hash of the model file(s). */ - cfg_scale?: number | number[]; + hash: string; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - denoising_start?: number; + path: string; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * File Size + * @description The size of the model in bytes. */ - denoising_end?: number; + file_size: number; /** - * Scheduler - * @description Scheduler to use during inference - * @default euler - * @enum {string} + * Name + * @description Name of the model. */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + name: string; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Description + * @description Model description */ - unet?: components["schemas"]["UNetField"] | null; + description: string | null; /** - * Control - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * IP-Adapter - * @description IP-Adapter to apply - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + source_api_response: string | null; /** - * T2I-Adapter - * @description T2I-Adapter(s) to apply - * @default null + * Cover Image + * @description Url for image to preview model */ - t2i_adapter?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + cover_image: string | null; /** - * CFG Rescale Multiplier - * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR - * @default 0 + * Config Path + * @description Path to the config for this model, if any. */ - cfg_rescale_multiplier?: number; + config_path: string | null; /** - * @description Latents tensor - * @default null + * Type + * @default controlnet + * @constant */ - latents?: components["schemas"]["LatentsField"] | null; + type: "controlnet"; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. - * @default null + * Format + * @default checkpoint + * @constant */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + format: "checkpoint"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * type - * @default denoise_latents + * Base + * @default flux * @constant */ - type: "denoise_latents"; + base: "flux"; }; - /** Denoise - SD1.5, SDXL + Metadata */ - DenoiseLatentsMetaInvocation: { + /** ControlNet_Checkpoint_SD1_Config */ + ControlNet_Checkpoint_SD1_Config: { /** - * @description Optional metadata to be saved with the image - * @default null + * Key + * @description A unique key for this model. */ - metadata?: components["schemas"]["MetadataField"] | null; + key: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Hash + * @description The hash of the model file(s). */ - id: string; + hash: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - is_intermediate?: boolean; + path: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * File Size + * @description The size of the model in bytes. */ - use_cache?: boolean; + file_size: number; /** - * Positive Conditioning - * @description Positive conditioning tensor - * @default null + * Name + * @description Name of the model. */ - positive_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; + name: string; /** - * Negative Conditioning - * @description Negative conditioning tensor - * @default null + * Description + * @description Model description */ - negative_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; + description: string | null; /** - * @description Noise tensor - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - noise?: components["schemas"]["LatentsField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Steps - * @description Number of steps to run - * @default 10 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - steps?: number; + source_api_response: string | null; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 7.5 + * Cover Image + * @description Url for image to preview model */ - cfg_scale?: number | number[]; + cover_image: string | null; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Config Path + * @description Path to the config for this model, if any. */ - denoising_start?: number; + config_path: string | null; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * Type + * @default controlnet + * @constant */ - denoising_end?: number; + type: "controlnet"; /** - * Scheduler - * @description Scheduler to use during inference - * @default euler - * @enum {string} + * Format + * @default checkpoint + * @constant */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + format: "checkpoint"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Base + * @default sd-1 + * @constant */ - unet?: components["schemas"]["UNetField"] | null; + base: "sd-1"; + }; + /** ControlNet_Checkpoint_SD2_Config */ + ControlNet_Checkpoint_SD2_Config: { /** - * Control - * @default null + * Key + * @description A unique key for this model. */ - control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + key: string; /** - * IP-Adapter - * @description IP-Adapter to apply - * @default null + * Hash + * @description The hash of the model file(s). */ - ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + hash: string; /** - * T2I-Adapter - * @description T2I-Adapter(s) to apply - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - t2i_adapter?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + path: string; /** - * CFG Rescale Multiplier - * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR - * @default 0 + * File Size + * @description The size of the model in bytes. */ - cfg_rescale_multiplier?: number; + file_size: number; /** - * @description Latents tensor - * @default null + * Name + * @description Name of the model. */ - latents?: components["schemas"]["LatentsField"] | null; + name: string; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. - * @default null + * Description + * @description Model description */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + description: string | null; /** - * type - * @default denoise_latents_meta - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "denoise_latents_meta"; - }; - /** - * DenoiseMaskField - * @description An inpaint mask field - */ - DenoiseMaskField: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Mask Name - * @description The name of the mask image + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - mask_name: string; + source_api_response: string | null; /** - * Masked Latents Name - * @description The name of the masked image latents - * @default null + * Cover Image + * @description Url for image to preview model */ - masked_latents_name?: string | null; + cover_image: string | null; /** - * Gradient - * @description Used for gradient inpainting - * @default false + * Config Path + * @description Path to the config for this model, if any. */ - gradient?: boolean; - }; - /** - * DenoiseMaskOutput - * @description Base class for nodes that output a single image - */ - DenoiseMaskOutput: { - /** @description Mask for denoise model run */ - denoise_mask: components["schemas"]["DenoiseMaskField"]; + config_path: string | null; /** - * type - * @default denoise_mask_output + * Type + * @default controlnet * @constant */ - type: "denoise_mask_output"; - }; - /** - * Depth Anything Depth Estimation - * @description Generates a depth map using a Depth Anything model. - */ - DepthAnythingDepthEstimationInvocation: { + type: "controlnet"; /** - * @description The board to save the image to - * @default null + * Format + * @default checkpoint + * @constant */ - board?: components["schemas"]["BoardField"] | null; + format: "checkpoint"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Base + * @default sd-2 + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + base: "sd-2"; + }; + /** ControlNet_Checkpoint_SDXL_Config */ + ControlNet_Checkpoint_SDXL_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * @description The image to process - * @default null + * File Size + * @description The size of the model in bytes. */ - image?: components["schemas"]["ImageField"] | null; + file_size: number; /** - * Model Size - * @description The size of the depth model to use - * @default small_v2 - * @enum {string} + * Name + * @description Name of the model. */ - model_size?: "large" | "base" | "small" | "small_v2"; + name: string; /** - * type - * @default depth_anything_depth_estimation - * @constant + * Description + * @description Model description */ - type: "depth_anything_depth_estimation"; - }; - /** - * Divide Integers - * @description Divides two numbers - */ - DivideInvocation: { + description: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source + * @description The original source of the model (path, URL or repo_id). */ - id: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - is_intermediate?: boolean; + source_api_response: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Cover Image + * @description Url for image to preview model */ - use_cache?: boolean; + cover_image: string | null; /** - * A - * @description The first number - * @default 0 + * Config Path + * @description Path to the config for this model, if any. */ - a?: number; - /** - * B - * @description The second number - * @default 0 - */ - b?: number; + config_path: string | null; /** - * type - * @default div + * Type + * @default controlnet * @constant */ - type: "div"; - }; - /** - * DownloadCancelledEvent - * @description Event model for download_cancelled - */ - DownloadCancelledEvent: { + type: "controlnet"; /** - * Timestamp - * @description The timestamp of the event + * Format + * @default checkpoint + * @constant */ - timestamp: number; + format: "checkpoint"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Source - * @description The source of the download + * Base + * @default sdxl + * @constant */ - source: string; + base: "sdxl"; }; /** - * DownloadCompleteEvent - * @description Event model for download_complete + * ControlNet_Checkpoint_ZImage_Config + * @description Model config for Z-Image Control adapter models (Safetensors checkpoint). + * + * Z-Image Control models are standalone adapters containing only the control layers + * (control_layers, control_all_x_embedder, control_noise_refiner) that extend + * the base Z-Image transformer with spatial conditioning capabilities. + * + * Supports: Canny, HED, Depth, Pose, MLSD. + * Recommended control_context_scale: 0.65-0.80. */ - DownloadCompleteEvent: { + ControlNet_Checkpoint_ZImage_Config: { /** - * Timestamp - * @description The timestamp of the event + * Key + * @description A unique key for this model. */ - timestamp: number; + key: string; /** - * Source - * @description The source of the download + * Hash + * @description The hash of the model file(s). */ - source: string; + hash: string; /** - * Download Path - * @description The local path where the download is saved + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - download_path: string; + path: string; /** - * Total Bytes - * @description The total number of bytes downloaded + * File Size + * @description The size of the model in bytes. */ - total_bytes: number; - }; - /** - * DownloadErrorEvent - * @description Event model for download_error - */ - DownloadErrorEvent: { + file_size: number; /** - * Timestamp - * @description The timestamp of the event + * Name + * @description Name of the model. */ - timestamp: number; + name: string; + /** + * Description + * @description Model description + */ + description: string | null; /** * Source - * @description The source of the download + * @description The original source of the model (path, URL or repo_id). */ source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Error Type - * @description The type of error + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - error_type: string; + source_api_response: string | null; /** - * Error - * @description The error message + * Cover Image + * @description Url for image to preview model */ - error: string; - }; - /** - * DownloadJob - * @description Class to monitor and control a model download request. - */ - DownloadJob: { + cover_image: string | null; /** - * Id - * @description Numeric ID of this job - * @default -1 + * Config Path + * @description Path to the config for this model, if any. */ - id?: number; + config_path: string | null; /** - * Dest - * Format: path - * @description Initial destination of downloaded model on local disk; a directory or file path + * Type + * @default controlnet + * @constant */ - dest: string; + type: "controlnet"; /** - * Download Path - * @description Final location of downloaded file or directory + * Format + * @default checkpoint + * @constant */ - download_path?: string | null; + format: "checkpoint"; /** - * @description Status of the download - * @default waiting + * Base + * @default z-image + * @constant */ - status?: components["schemas"]["DownloadJobStatus"]; + base: "z-image"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + }; + /** ControlNet_Diffusers_FLUX_Config */ + ControlNet_Diffusers_FLUX_Config: { /** - * Bytes - * @description Bytes downloaded so far - * @default 0 + * Key + * @description A unique key for this model. */ - bytes?: number; + key: string; /** - * Total Bytes - * @description Total file size (bytes) - * @default 0 + * Hash + * @description The hash of the model file(s). */ - total_bytes?: number; + hash: string; /** - * Error Type - * @description Name of exception that caused an error + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - error_type?: string | null; + path: string; /** - * Error - * @description Traceback of the exception that caused an error + * File Size + * @description The size of the model in bytes. */ - error?: string | null; + file_size: number; /** - * Source - * Format: uri - * @description Where to download from. Specific types specified in child classes. + * Name + * @description Name of the model. */ - source: string; + name: string; /** - * Access Token - * @description authorization token for protected resources + * Description + * @description Model description */ - access_token?: string | null; + description: string | null; /** - * Priority - * @description Queue priority; lower values are higher priority - * @default 10 + * Source + * @description The original source of the model (path, URL or repo_id). */ - priority?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Job Started - * @description Timestamp for when the download job started + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - job_started?: string | null; + source_api_response: string | null; /** - * Job Ended - * @description Timestamp for when the download job ende1d (completed or errored) + * Cover Image + * @description Url for image to preview model */ - job_ended?: string | null; + cover_image: string | null; /** - * Content Type - * @description Content type of downloaded file + * Format + * @default diffusers + * @constant */ - content_type?: string | null; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Canonical Url - * @description Canonical URL to request on resume + * Type + * @default controlnet + * @constant */ - canonical_url?: string | null; + type: "controlnet"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Etag - * @description ETag from the remote server, if available + * Base + * @default flux + * @constant */ - etag?: string | null; + base: "flux"; + }; + /** ControlNet_Diffusers_SD1_Config */ + ControlNet_Diffusers_SD1_Config: { /** - * Last Modified - * @description Last-Modified from the remote server, if available + * Key + * @description A unique key for this model. */ - last_modified?: string | null; + key: string; /** - * Final Url - * @description Final resolved URL after redirects, if available + * Hash + * @description The hash of the model file(s). */ - final_url?: string | null; + hash: string; /** - * Expected Total Bytes - * @description Expected total size of the download + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - expected_total_bytes?: number | null; + path: string; /** - * Resume Required - * @description True if server refused resume; restart required - * @default false + * File Size + * @description The size of the model in bytes. */ - resume_required?: boolean; + file_size: number; /** - * Resume Message - * @description Message explaining why resume is required - */ - resume_message?: string | null; - /** - * Resume From Scratch - * @description True if resume metadata existed but the partial file was missing and the download restarted from the beginning - * @default false + * Name + * @description Name of the model. */ - resume_from_scratch?: boolean; - }; - /** - * DownloadJobStatus - * @description State of a download job. - * @enum {string} - */ - DownloadJobStatus: "waiting" | "running" | "paused" | "completed" | "cancelled" | "error"; - /** - * DownloadPausedEvent - * @description Event model for download_paused - */ - DownloadPausedEvent: { + name: string; /** - * Timestamp - * @description The timestamp of the event + * Description + * @description Model description */ - timestamp: number; + description: string | null; /** * Source - * @description The source of the download + * @description The original source of the model (path, URL or repo_id). */ source: string; - }; - /** - * DownloadProgressEvent - * @description Event model for download_progress - */ - DownloadProgressEvent: { + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Timestamp - * @description The timestamp of the event + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - timestamp: number; + source_api_response: string | null; /** - * Source - * @description The source of the download + * Cover Image + * @description Url for image to preview model */ - source: string; + cover_image: string | null; /** - * Download Path - * @description The local path where the download is saved + * Format + * @default diffusers + * @constant */ - download_path: string; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Current Bytes - * @description The number of bytes downloaded so far + * Type + * @default controlnet + * @constant */ - current_bytes: number; + type: "controlnet"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Total Bytes - * @description The total number of bytes to be downloaded + * Base + * @default sd-1 + * @constant */ - total_bytes: number; + base: "sd-1"; }; - /** - * DownloadStartedEvent - * @description Event model for download_started - */ - DownloadStartedEvent: { + /** ControlNet_Diffusers_SD2_Config */ + ControlNet_Diffusers_SD2_Config: { /** - * Timestamp - * @description The timestamp of the event + * Key + * @description A unique key for this model. */ - timestamp: number; + key: string; /** - * Source - * @description The source of the download + * Hash + * @description The hash of the model file(s). */ - source: string; + hash: string; /** - * Download Path - * @description The local path where the download is saved + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - download_path: string; - }; - /** - * Dynamic Prompt - * @description Parses a prompt using adieyal/dynamicprompts' random or combinatorial generator - */ - DynamicPromptInvocation: { + path: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * File Size + * @description The size of the model in bytes. */ - id: string; + file_size: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Name + * @description Name of the model. */ - is_intermediate?: boolean; + name: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Description + * @description Model description */ - use_cache?: boolean; + description: string | null; /** - * Prompt - * @description The prompt to parse with dynamicprompts - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - prompt?: string | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Max Prompts - * @description The number of prompts to generate - * @default 1 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - max_prompts?: number; + source_api_response: string | null; /** - * Combinatorial - * @description Whether to use the combinatorial generator - * @default false + * Cover Image + * @description Url for image to preview model */ - combinatorial?: boolean; + cover_image: string | null; /** - * type - * @default dynamic_prompt + * Format + * @default diffusers * @constant */ - type: "dynamic_prompt"; - }; - /** DynamicPromptsResponse */ - DynamicPromptsResponse: { - /** Prompts */ - prompts: string[]; - /** Error */ - error?: string | null; - }; - /** - * Upscale (RealESRGAN) - * @description Upscales an image using RealESRGAN. - */ - ESRGANInvocation: { + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * @description The board to save the image to - * @default null + * Type + * @default controlnet + * @constant */ - board?: components["schemas"]["BoardField"] | null; + type: "controlnet"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Base + * @default sd-2 + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + base: "sd-2"; + }; + /** ControlNet_Diffusers_SDXL_Config */ + ControlNet_Diffusers_SDXL_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * @description The input image - * @default null + * File Size + * @description The size of the model in bytes. */ - image?: components["schemas"]["ImageField"] | null; + file_size: number; /** - * Model Name - * @description The Real-ESRGAN model to use - * @default RealESRGAN_x4plus.pth - * @enum {string} + * Name + * @description Name of the model. */ - model_name?: "RealESRGAN_x4plus.pth" | "RealESRGAN_x4plus_anime_6B.pth" | "ESRGAN_SRx4_DF2KOST_official-ff704c30.pth" | "RealESRGAN_x2plus.pth"; + name: string; /** - * Tile Size - * @description Tile size for tiled ESRGAN upscaling (0=tiling disabled) - * @default 400 + * Description + * @description Model description */ - tile_size?: number; + description: string | null; /** - * type - * @default esrgan - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "esrgan"; - }; - /** Edge */ - Edge: { - /** @description The connection for the edge's from node and field */ - source: components["schemas"]["EdgeConnection"]; - /** @description The connection for the edge's to node and field */ - destination: components["schemas"]["EdgeConnection"]; - }; - /** EdgeConnection */ - EdgeConnection: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Node Id - * @description The id of the node for this edge connection + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - node_id: string; + source_api_response: string | null; /** - * Field - * @description The field for this connection - */ - field: string; - }; - /** EnqueueBatchResult */ - EnqueueBatchResult: { - /** - * Queue Id - * @description The ID of the queue - */ - queue_id: string; - /** - * Enqueued - * @description The total number of queue items enqueued + * Cover Image + * @description Url for image to preview model */ - enqueued: number; + cover_image: string | null; /** - * Requested - * @description The total number of queue items requested to be enqueued + * Format + * @default diffusers + * @constant */ - requested: number; - /** @description The batch that was enqueued */ - batch: components["schemas"]["Batch"]; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Priority - * @description The priority of the enqueued batch + * Type + * @default controlnet + * @constant */ - priority: number; + type: "controlnet"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Item Ids - * @description The IDs of the queue items that were enqueued + * Base + * @default sdxl + * @constant */ - item_ids: number[]; + base: "sdxl"; }; /** - * Expand Mask with Fade - * @description Expands a mask with a fade effect. The mask uses black to indicate areas to keep from the generated image and white for areas to discard. - * The mask is thresholded to create a binary mask, and then a distance transform is applied to create a fade effect. - * The fade size is specified in pixels, and the mask is expanded by that amount. The result is a mask with a smooth transition from black to white. - * If the fade size is 0, the mask is returned as-is. + * ControlOutput + * @description node output for ControlNet info */ - ExpandMaskWithFadeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + ControlOutput: { + /** @description ControlNet(s) to apply */ + control: components["schemas"]["ControlField"]; /** - * @description Optional metadata to be saved with the image - * @default null + * type + * @default control_output + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "control_output"; + }; + /** + * Coordinated Noise (Flux) + * @description Generates latent noise that is stable for the given coordinates. + * + * That is, the noise at channel=1 x=3 y=4 for seed=42 will always be the same, regardless of the + * total size (width and height) of the region. + * + * It makes reproducible noise that you're able to crop or scroll. + * + * This variant has 16 channels for Flux. + */ + CoordinatedFluxNoiseInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -7983,41 +8839,65 @@ export type components = { */ use_cache?: boolean; /** - * @description The mask to expand + * Seed + * @description Seed for random number generation * @default null */ - mask?: components["schemas"]["ImageField"] | null; + seed?: number | null; /** - * Threshold - * @description The threshold for the binary mask (0-255) + * Width + * @description Width of output (px) + * @default 512 + */ + width?: number; + /** + * Height + * @description Height of output (px) + * @default 512 + */ + height?: number; + /** + * X Offset + * @description x-coordinate of the lower edge * @default 0 */ - threshold?: number; + x_offset?: number; /** - * Fade Size Px - * @description The size of the fade in pixels - * @default 32 + * Y Offset + * @description y-coordinate of the lower edge + * @default 0 */ - fade_size_px?: number; + y_offset?: number; + /** + * Channel Offset + * @description coordinate of the first channel + * @default 0 + */ + channel_offset?: number; + /** + * Algorithm + * @description psuedo-random generation algorithm + * @default pcg64 + * @enum {string} + */ + algorithm?: "pcg64" | "morton+farmhash"; /** * type - * @default expand_mask_with_fade + * @default noise_coordinated_flux * @constant */ - type: "expand_mask_with_fade"; - }; - /** ExposedField */ - ExposedField: { - /** Nodeid */ - nodeId: string; - /** Fieldname */ - fieldName: string; + type: "noise_coordinated_flux"; }; /** - * Apply LoRA Collection - FLUX - * @description Applies a collection of LoRAs to a FLUX transformer. + * Coordinated Noise + * @description Generates latent noise that is stable for the given coordinates. + * + * That is, the noise at channel=1 x=3 y=4 for seed=42 will always be the same, regardless of the + * total size (width and height) of the region. + * + * It makes reproducible noise that you're able to crop or scroll. */ - FLUXLoRACollectionLoader: { + CoordinatedNoiseInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8036,122 +8916,60 @@ export type components = { */ use_cache?: boolean; /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. - * @default null - */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; - /** - * Transformer - * @description Transformer - * @default null - */ - transformer?: components["schemas"]["TransformerField"] | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip?: components["schemas"]["CLIPField"] | null; - /** - * T5 Encoder - * @description T5 tokenizer and text encoder + * Seed + * @description Seed for random number generation * @default null */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; - /** - * type - * @default flux_lora_collection_loader - * @constant - */ - type: "flux_lora_collection_loader"; - }; - /** - * FLUXRedux_Checkpoint_Config - * @description Model config for FLUX Tools Redux model. - */ - FLUXRedux_Checkpoint_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; + seed?: number | null; /** - * Description - * @description Model description + * Width + * @description Width of output (px) + * @default 512 */ - description: string | null; + width?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Height + * @description Height of output (px) + * @default 512 */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + height?: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * X Offset + * @description x-coordinate of the lower edge + * @default 0 */ - source_api_response: string | null; + x_offset?: number; /** - * Cover Image - * @description Url for image to preview model + * Y Offset + * @description y-coordinate of the lower edge + * @default 0 */ - cover_image: string | null; + y_offset?: number; /** - * Type - * @default flux_redux - * @constant + * Channel Offset + * @description coordinate of the first channel + * @default 0 */ - type: "flux_redux"; + channel_offset?: number; /** - * Format - * @default checkpoint - * @constant + * Algorithm + * @description psuedo-random generation algorithm + * @default pcg64 + * @enum {string} */ - format: "checkpoint"; + algorithm?: "pcg64" | "morton+farmhash"; /** - * Base - * @default flux + * type + * @default noise_coordinated * @constant */ - base: "flux"; + type: "noise_coordinated"; }; /** - * FaceIdentifier - * @description Outputs an image with detected face IDs printed on each face. For use with other FaceTools. + * Core Metadata + * @description Used internally by Invoke to collect metadata for generations. */ - FaceIdentifierInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + CoreMetadataInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8170,264 +8988,225 @@ export type components = { */ use_cache?: boolean; /** - * @description Image to face detect + * Generation Mode + * @description The generation mode that output this image * @default null */ - image?: components["schemas"]["ImageField"] | null; + generation_mode?: ("txt2img" | "img2img" | "inpaint" | "outpaint" | "sdxl_txt2img" | "sdxl_img2img" | "sdxl_inpaint" | "sdxl_outpaint" | "flux_txt2img" | "flux_img2img" | "flux_inpaint" | "flux_outpaint" | "flux2_txt2img" | "flux2_img2img" | "flux2_inpaint" | "flux2_outpaint" | "sd3_txt2img" | "sd3_img2img" | "sd3_inpaint" | "sd3_outpaint" | "cogview4_txt2img" | "cogview4_img2img" | "cogview4_inpaint" | "cogview4_outpaint" | "z_image_txt2img" | "z_image_img2img" | "z_image_inpaint" | "z_image_outpaint") | null; /** - * Minimum Confidence - * @description Minimum confidence for face detection (lower if detection is failing) - * @default 0.5 + * Positive Prompt + * @description The positive prompt parameter + * @default null */ - minimum_confidence?: number; + positive_prompt?: string | null; /** - * Chunk - * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. - * @default false + * Negative Prompt + * @description The negative prompt parameter + * @default null */ - chunk?: boolean; + negative_prompt?: string | null; /** - * type - * @default face_identifier - * @constant + * Width + * @description The width parameter + * @default null */ - type: "face_identifier"; - }; - /** - * FaceMask - * @description Face mask creation using mediapipe face detection - */ - FaceMaskInvocation: { + width?: number | null; /** - * @description Optional metadata to be saved with the image + * Height + * @description The height parameter * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + height?: number | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Seed + * @description The seed used for noise generation + * @default null */ - id: string; + seed?: number | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Rand Device + * @description The device used for random number generation + * @default null */ - is_intermediate?: boolean; + rand_device?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Cfg Scale + * @description The classifier-free guidance scale parameter + * @default null */ - use_cache?: boolean; + cfg_scale?: number | null; /** - * @description Image to face detect + * Cfg Rescale Multiplier + * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR * @default null */ - image?: components["schemas"]["ImageField"] | null; + cfg_rescale_multiplier?: number | null; /** - * Face Ids - * @description Comma-separated list of face ids to mask eg '0,2,7'. Numbered from 0. Leave empty to mask all. Find face IDs with FaceIdentifier node. - * @default + * Steps + * @description The number of steps used for inference + * @default null */ - face_ids?: string; + steps?: number | null; /** - * Minimum Confidence - * @description Minimum confidence for face detection (lower if detection is failing) - * @default 0.5 + * Scheduler + * @description The scheduler used for inference + * @default null */ - minimum_confidence?: number; + scheduler?: string | null; /** - * X Offset - * @description Offset for the X-axis of the face mask - * @default 0 + * Seamless X + * @description Whether seamless tiling was used on the X axis + * @default null */ - x_offset?: number; + seamless_x?: boolean | null; /** - * Y Offset - * @description Offset for the Y-axis of the face mask - * @default 0 + * Seamless Y + * @description Whether seamless tiling was used on the Y axis + * @default null */ - y_offset?: number; + seamless_y?: boolean | null; /** - * Chunk - * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. - * @default false + * Clip Skip + * @description The number of skipped CLIP layers + * @default null */ - chunk?: boolean; + clip_skip?: number | null; /** - * Invert Mask - * @description Toggle to invert the mask - * @default false + * @description The main model used for inference + * @default null */ - invert_mask?: boolean; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * type - * @default face_mask_detection - * @constant + * Controlnets + * @description The ControlNets used for inference + * @default null */ - type: "face_mask_detection"; - }; - /** - * FaceMaskOutput - * @description Base class for FaceMask output - */ - FaceMaskOutput: { - /** @description The output image */ - image: components["schemas"]["ImageField"]; + controlnets?: components["schemas"]["ControlNetMetadataField"][] | null; /** - * Width - * @description The width of the image in pixels + * Ipadapters + * @description The IP Adapters used for inference + * @default null */ - width: number; + ipAdapters?: components["schemas"]["IPAdapterMetadataField"][] | null; /** - * Height - * @description The height of the image in pixels + * T2Iadapters + * @description The IP Adapters used for inference + * @default null */ - height: number; + t2iAdapters?: components["schemas"]["T2IAdapterMetadataField"][] | null; /** - * type - * @default face_mask_output - * @constant + * Loras + * @description The LoRAs used for inference + * @default null */ - type: "face_mask_output"; - /** @description The output mask */ - mask: components["schemas"]["ImageField"]; - }; - /** - * FaceOff - * @description Bound, extract, and mask a face from an image using MediaPipe detection - */ - FaceOffInvocation: { + loras?: components["schemas"]["LoRAMetadataField"][] | null; /** - * @description Optional metadata to be saved with the image + * Strength + * @description The strength used for latents-to-latents * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + strength?: number | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Init Image + * @description The name of the initial image + * @default null */ - id: string; + init_image?: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * @description The VAE used for decoding, if the main model's default was not used + * @default null */ - is_intermediate?: boolean; + vae?: components["schemas"]["ModelIdentifierField"] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * @description The Qwen3 text encoder model used for Z-Image inference + * @default null */ - use_cache?: boolean; + qwen3_encoder?: components["schemas"]["ModelIdentifierField"] | null; /** - * @description Image for face detection + * Hrf Enabled + * @description Whether or not high resolution fix was enabled. * @default null */ - image?: components["schemas"]["ImageField"] | null; + hrf_enabled?: boolean | null; /** - * Face Id - * @description The face ID to process, numbered from 0. Multiple faces not supported. Find a face's ID with FaceIdentifier node. - * @default 0 + * Hrf Method + * @description The high resolution fix upscale method. + * @default null */ - face_id?: number; + hrf_method?: string | null; /** - * Minimum Confidence - * @description Minimum confidence for face detection (lower if detection is failing) - * @default 0.5 + * Hrf Strength + * @description The high resolution fix img2img strength used in the upscale pass. + * @default null */ - minimum_confidence?: number; + hrf_strength?: number | null; /** - * X Offset - * @description X-axis offset of the mask - * @default 0 + * Positive Style Prompt + * @description The positive style prompt parameter + * @default null */ - x_offset?: number; + positive_style_prompt?: string | null; /** - * Y Offset - * @description Y-axis offset of the mask - * @default 0 + * Negative Style Prompt + * @description The negative style prompt parameter + * @default null */ - y_offset?: number; + negative_style_prompt?: string | null; /** - * Padding - * @description All-axis padding around the mask in pixels - * @default 0 + * @description The SDXL Refiner model used + * @default null */ - padding?: number; + refiner_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Chunk - * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. - * @default false + * Refiner Cfg Scale + * @description The classifier-free guidance scale parameter used for the refiner + * @default null */ - chunk?: boolean; + refiner_cfg_scale?: number | null; /** - * type - * @default face_off - * @constant + * Refiner Steps + * @description The number of steps used for the refiner + * @default null */ - type: "face_off"; - }; - /** - * FaceOffOutput - * @description Base class for FaceOff Output - */ - FaceOffOutput: { - /** @description The output image */ - image: components["schemas"]["ImageField"]; + refiner_steps?: number | null; /** - * Width - * @description The width of the image in pixels + * Refiner Scheduler + * @description The scheduler used for the refiner + * @default null */ - width: number; + refiner_scheduler?: string | null; /** - * Height - * @description The height of the image in pixels + * Refiner Positive Aesthetic Score + * @description The aesthetic score used for the refiner + * @default null */ - height: number; + refiner_positive_aesthetic_score?: number | null; /** - * type - * @default face_off_output - * @constant + * Refiner Negative Aesthetic Score + * @description The aesthetic score used for the refiner + * @default null */ - type: "face_off_output"; - /** @description The output mask */ - mask: components["schemas"]["ImageField"]; + refiner_negative_aesthetic_score?: number | null; /** - * X - * @description The x coordinate of the bounding box's left side + * Refiner Start + * @description The start value used for refiner denoising + * @default null */ - x: number; + refiner_start?: number | null; /** - * Y - * @description The y coordinate of the bounding box's top side + * type + * @default core_metadata + * @constant */ - y: number; + type: "core_metadata"; + } & { + [key: string]: unknown; }; /** - * FieldKind - * @description The kind of field. - * - `Input`: An input field on a node. - * - `Output`: An output field on a node. - * - `Internal`: A field which is treated as an input, but cannot be used in node definitions. Metadata is - * one example. It is provided to nodes via the WithMetadata class, and we want to reserve the field name - * "metadata" for this on all nodes. `FieldKind` is used to short-circuit the field name validation logic, - * allowing "metadata" for that field. - * - `NodeAttribute`: The field is a node attribute. These are fields which are not inputs or outputs, - * but which are used to store information about the node. For example, the `id` and `type` fields are node - * attributes. - * - * The presence of this in `json_schema_extra["field_kind"]` is used when initializing node schemas on app - * startup, and when generating the OpenAPI schema for the workflow editor. - * @enum {string} - */ - FieldKind: "input" | "output" | "internal" | "node_attribute"; - /** - * Float Batch - * @description Create a batched generation, where the workflow is executed once for each float in the batch. + * Create Denoise Mask + * @description Creates mask for denoising model run. */ - FloatBatchInvocation: { + CreateDenoiseMaskInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8446,30 +9225,44 @@ export type components = { */ use_cache?: boolean; /** - * Batch Group - * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. - * @default None - * @enum {string} + * @description VAE + * @default null */ - batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; + vae?: components["schemas"]["VAEField"] | null; /** - * Floats - * @description The floats to batch over + * @description Image which will be masked * @default null */ - floats?: number[] | null; + image?: components["schemas"]["ImageField"] | null; + /** + * @description The mask to use when pasting + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * Tiled + * @description Processing using overlapping tiles (reduce memory consumption) + * @default false + */ + tiled?: boolean; + /** + * Fp32 + * @description Whether or not to use full float32 precision + * @default false + */ + fp32?: boolean; /** * type - * @default float_batch + * @default create_denoise_mask * @constant */ - type: "float_batch"; + type: "create_denoise_mask"; }; /** - * Float Collection Primitive - * @description A collection of float primitive values + * Create Gradient Mask + * @description Creates mask for denoising. */ - FloatCollectionInvocation: { + CreateGradientMaskInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8488,40 +9281,80 @@ export type components = { */ use_cache?: boolean; /** - * Collection - * @description The collection of float values - * @default [] + * @description Image which will be masked + * @default null */ - collection?: number[]; + mask?: components["schemas"]["ImageField"] | null; /** - * type - * @default float_collection - * @constant + * Edge Radius + * @description How far to expand the edges of the mask + * @default 16 */ - type: "float_collection"; - }; - /** - * FloatCollectionOutput - * @description Base class for nodes that output a collection of floats - */ - FloatCollectionOutput: { + edge_radius?: number; /** - * Collection - * @description The float collection + * Coherence Mode + * @default Gaussian Blur + * @enum {string} */ - collection: number[]; + coherence_mode?: "Gaussian Blur" | "Box Blur" | "Staged"; + /** + * Minimum Denoise + * @description Minimum denoise level for the coherence region + * @default 0 + */ + minimum_denoise?: number; + /** + * [OPTIONAL] Image + * @description OPTIONAL: Only connect for specialized Inpainting models, masked_latents will be generated from the image with the VAE + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * [OPTIONAL] UNet + * @description OPTIONAL: If the Unet is a specialized Inpainting model, masked_latents will be generated from the image with the VAE + * @default null + */ + unet?: components["schemas"]["UNetField"] | null; + /** + * [OPTIONAL] VAE + * @description OPTIONAL: Only connect for specialized Inpainting models, masked_latents will be generated from the image with the VAE + * @default null + */ + vae?: components["schemas"]["VAEField"] | null; + /** + * Tiled + * @description Processing using overlapping tiles (reduce memory consumption) + * @default false + */ + tiled?: boolean; + /** + * Fp32 + * @description Whether or not to use full float32 precision + * @default false + */ + fp32?: boolean; /** * type - * @default float_collection_output + * @default create_gradient_mask * @constant */ - type: "float_collection_output"; + type: "create_gradient_mask"; }; /** - * Float Generator - * @description Generated a range of floats for use in a batched generation + * Crop Image to Bounding Box + * @description Crop an image to the given bounding box. If the bounding box is omitted, the image is cropped to the non-transparent pixels. */ - FloatGenerator: { + CropImageToBoundingBoxInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8540,41 +9373,28 @@ export type components = { */ use_cache?: boolean; /** - * Generator Type - * @description The float generator. - */ - generator: components["schemas"]["FloatGeneratorField"]; - /** - * type - * @default float_generator - * @constant + * @description The image to crop + * @default null */ - type: "float_generator"; - }; - /** FloatGeneratorField */ - FloatGeneratorField: Record; - /** - * FloatGeneratorOutput - * @description Base class for nodes that output a collection of floats - */ - FloatGeneratorOutput: { + image?: components["schemas"]["ImageField"] | null; /** - * Floats - * @description The generated floats + * @description The bounding box to crop the image to + * @default null */ - floats: number[]; + bounding_box?: components["schemas"]["BoundingBoxField"] | null; /** * type - * @default float_generator_output + * @default crop_image_to_bounding_box * @constant */ - type: "float_generator_output"; + type: "crop_image_to_bounding_box"; }; /** - * Float Primitive - * @description A float primitive value + * Crop Latents + * @description Crops a latent-space tensor to a box specified in image-space. The box dimensions and coordinates must be + * divisible by the latent scale factor of 8. */ - FloatInvocation: { + CropLatentsCoreInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8593,23 +9413,46 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The float value - * @default 0 + * @description Latents tensor + * @default null */ - value?: number; + latents?: components["schemas"]["LatentsField"] | null; + /** + * X + * @description The left x coordinate (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @default null + */ + x?: number | null; + /** + * Y + * @description The top y coordinate (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @default null + */ + y?: number | null; + /** + * Width + * @description The width (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @default null + */ + width?: number | null; + /** + * Height + * @description The height (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @default null + */ + height?: number | null; /** * type - * @default float + * @default crop_latents * @constant */ - type: "float"; + type: "crop_latents"; }; /** - * Float Range - * @description Creates a range + * Crop Latents + * @description Crops latents */ - FloatLinearRangeInvocation: { + CropLatentsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8628,35 +9471,46 @@ export type components = { */ use_cache?: boolean; /** - * Start - * @description The first value of the range - * @default 5 + * @description Latents tensor + * @default null */ - start?: number; + latents?: components["schemas"]["LatentsField"] | null; /** - * Stop - * @description The last value of the range - * @default 10 + * Width + * @description Width of output (px) + * @default null */ - stop?: number; + width?: number | null; /** - * Steps - * @description number of values to interpolate over (including start and stop) - * @default 30 + * Height + * @description Width of output (px) + * @default null */ - steps?: number; + height?: number | null; + /** + * X Offset + * @description x-coordinate + * @default null + */ + x_offset?: number | null; + /** + * Y Offset + * @description y-coordinate + * @default null + */ + y_offset?: number | null; /** * type - * @default float_range + * @default lcrop * @constant */ - type: "float_range"; + type: "lcrop"; }; /** - * Float Math - * @description Performs floating point math. + * Crossover Prompt + * @description Performs a crossover operation on two parent seed vectors to generate a new prompt. */ - FloatMathInvocation: { + CrossoverPromptInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8671,57 +9525,87 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * Operation - * @description The operation to perform - * @default ADD - * @enum {string} + * Resolutions Dict + * @description Private field for id substitutions dict cache + * @default {} */ - operation?: "ADD" | "SUB" | "MUL" | "DIV" | "EXP" | "ABS" | "SQRT" | "MIN" | "MAX"; + resolutions_dict?: { + [key: string]: unknown; + }; /** - * A - * @description The first number - * @default 1 + * Lookups + * @description Lookup table(s) containing template(s) (JSON) + * @default [] */ - a?: number; + lookups?: string | string[]; /** - * B - * @description The second number - * @default 1 + * Remove Negatives + * @description Whether to strip out text between [] + * @default false */ - b?: number; + remove_negatives?: boolean; /** - * type - * @default float_math - * @constant + * Strip Parens Probability + * @description Probability of removing attention group weightings + * @default 0 */ - type: "float_math"; - }; - /** - * FloatOutput - * @description Base class for nodes that output a single float - */ - FloatOutput: { + strip_parens_probability?: number; /** - * Value - * @description The output float + * Resolutions + * @description JSON structure of substitutions by id by tag + * @default [] */ - value: number; + resolutions?: string | string[]; + /** + * Parent A Seed Vector In + * @description JSON array of seeds for Parent A's generation + * @default null + */ + parent_a_seed_vector_in?: string | null; + /** + * Parent B Seed Vector In + * @description JSON array of seeds for Parent B's generation + * @default null + */ + parent_b_seed_vector_in?: string | null; + /** + * Child A Or B + * @description True for Child A (Parent A + B's branch), False for Child B (Parent B + A's branch) + * @default true + */ + child_a_or_b?: boolean; + /** + * Crossover Non Terminal + * @description Optional: The non-terminal (key in lookups) to target for the crossover branch. If None, a random one will be chosen from available non-terminals. + * @default null + */ + crossover_non_terminal?: string | null; /** * type - * @default float_output + * @default crossover_prompt * @constant */ - type: "float_output"; + type: "crossover_prompt"; }; /** - * Float To Integer - * @description Rounds a float number to (a multiple of) an integer. + * OpenCV Inpaint + * @description Simple inpaint using opencv. */ - FloatToIntegerInvocation: { + CvInpaintInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8740,39 +9624,37 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The value to round - * @default 0 - */ - value?: number; - /** - * Multiple of - * @description The multiple to round to - * @default 1 + * @description The image to inpaint + * @default null */ - multiple?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Method - * @description The method to use for rounding - * @default Nearest - * @enum {string} + * @description The mask to use when inpainting + * @default null */ - method?: "Nearest" | "Floor" | "Ceiling" | "Truncate"; + mask?: components["schemas"]["ImageField"] | null; /** * type - * @default float_to_int + * @default cv_inpaint * @constant */ - type: "float_to_int"; + type: "cv_inpaint"; }; /** - * FLUX2 Denoise - * @description Run denoising process with a FLUX.2 Klein transformer model. - * - * This node is designed for FLUX.2 Klein models which use Qwen3 as the text encoder. - * It does not support ControlNet, IP-Adapters, or regional prompting. + * DW Openpose Detection + * @description Generates an openpose pose from an image using DWPose */ - Flux2DenoiseInvocation: { + DWOpenposeDetectionInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8791,109 +9673,77 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @description The image to process * @default null */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; - /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 - */ - denoising_start?: number; - /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 - */ - denoising_end?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Add Noise - * @description Add noise based on denoising start. + * Draw Body * @default true */ - add_noise?: boolean; - /** - * Transformer - * @description Flux model (Transformer) to load - * @default null - */ - transformer?: components["schemas"]["TransformerField"] | null; - /** - * @description Positive conditioning tensor - * @default null - */ - positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | null; - /** - * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. - * @default null - */ - negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | null; + draw_body?: boolean; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 1 + * Draw Face + * @default false */ - cfg_scale?: number; + draw_face?: boolean; /** - * Width - * @description Width of the generated image. - * @default 1024 + * Draw Hands + * @default false */ - width?: number; + draw_hands?: boolean; /** - * Height - * @description Height of the generated image. - * @default 1024 + * type + * @default dw_openpose_detection + * @constant */ - height?: number; + type: "dw_openpose_detection"; + }; + /** + * Decode Invisible Watermark + * @description Decode an invisible watermark from an image. + */ + DecodeInvisibleWatermarkInvocation: { /** - * Num Steps - * @description Number of diffusion steps. Use 4 for distilled models, 28+ for base models. - * @default 4 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - num_steps?: number; + id: string; /** - * Scheduler - * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. - * @default euler - * @enum {string} + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - scheduler?: "euler" | "heun" | "lcm"; + is_intermediate?: boolean; /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - seed?: number; + use_cache?: boolean; /** - * @description FLUX.2 VAE model (required for BN statistics). + * @description The image to decode the watermark from * @default null */ - vae?: components["schemas"]["VAEField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Reference Images - * @description FLUX Kontext conditioning (reference images for multi-reference image editing). - * @default null + * Length + * @description The expected watermark length in bytes + * @default 8 */ - kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; + length?: number; /** * type - * @default flux2_denoise + * @default decode_watermark * @constant */ - type: "flux2_denoise"; + type: "decode_watermark"; }; /** - * Apply LoRA Collection - Flux2 Klein - * @description Applies a collection of LoRAs to a FLUX.2 Klein transformer and/or Qwen3 text encoder. + * Default XYImage Tile Generator + * @description Cuts up an image into overlapping tiles and outputs a string representation of the tiles to use */ - Flux2KleinLoRACollectionLoader: { + DefaultXYTileGenerator: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8912,121 +9762,128 @@ export type components = { */ use_cache?: boolean; /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. + * @description The input image * @default null */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Transformer - * @description Transformer - * @default null + * Tile Width + * @description x resolution of generation tile (must be a multiple of 8) + * @default 576 */ - transformer?: components["schemas"]["TransformerField"] | null; + tile_width?: number; /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder - * @default null + * Tile Height + * @description y resolution of generation tile (must be a multiple of 8) + * @default 576 */ - qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; + tile_height?: number; + /** + * Overlap + * @description tile overlap size (must be a multiple of 8) + * @default 128 + */ + overlap?: number; + /** + * Adjust Tile Size + * @description adjust tile size to account for overlap + * @default true + */ + adjust_tile_size?: boolean; /** * type - * @default flux2_klein_lora_collection_loader + * @default default_xy_tile_generator * @constant */ - type: "flux2_klein_lora_collection_loader"; + type: "default_xy_tile_generator"; }; /** - * Apply LoRA - Flux2 Klein - * @description Apply a LoRA model to a FLUX.2 Klein transformer and/or Qwen3 text encoder. + * DeleteAllExceptCurrentResult + * @description Result of deleting all except current */ - Flux2KleinLoRALoaderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + DeleteAllExceptCurrentResult: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Deleted + * @description Number of queue items deleted */ - is_intermediate?: boolean; + deleted: number; + }; + /** DeleteBoardResult */ + DeleteBoardResult: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Board Id + * @description The id of the board that was deleted. */ - use_cache?: boolean; + board_id: string; /** - * LoRA - * @description LoRA model to load - * @default null + * Deleted Board Images + * @description The image names of the board-images relationships that were deleted. */ - lora?: components["schemas"]["ModelIdentifierField"] | null; + deleted_board_images: string[]; /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 + * Deleted Images + * @description The names of the images that were deleted. */ - weight?: number; + deleted_images: string[]; + }; + /** + * DeleteByDestinationResult + * @description Result of deleting by a destination + */ + DeleteByDestinationResult: { /** - * Transformer - * @description Transformer - * @default null + * Deleted + * @description Number of queue items deleted */ - transformer?: components["schemas"]["TransformerField"] | null; + deleted: number; + }; + /** DeleteImagesResult */ + DeleteImagesResult: { /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder - * @default null + * Affected Boards + * @description The ids of boards affected by the delete operation */ - qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; + affected_boards: string[]; /** - * type - * @default flux2_klein_lora_loader - * @constant + * Deleted Images + * @description The names of the images that were deleted */ - type: "flux2_klein_lora_loader"; + deleted_images: string[]; }; /** - * Flux2KleinLoRALoaderOutput - * @description FLUX.2 Klein LoRA Loader Output + * DeleteOrphanedModelsRequest + * @description Request to delete specific orphaned model directories. */ - Flux2KleinLoRALoaderOutput: { + DeleteOrphanedModelsRequest: { /** - * Transformer - * @description Transformer - * @default null + * Paths + * @description List of relative paths to delete */ - transformer: components["schemas"]["TransformerField"] | null; + paths: string[]; + }; + /** + * DeleteOrphanedModelsResponse + * @description Response from deleting orphaned models. + */ + DeleteOrphanedModelsResponse: { /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder - * @default null + * Deleted + * @description Paths that were successfully deleted */ - qwen3_encoder: components["schemas"]["Qwen3EncoderField"] | null; + deleted: string[]; /** - * type - * @default flux2_klein_lora_loader_output - * @constant + * Errors + * @description Paths that had errors, with error messages */ - type: "flux2_klein_lora_loader_output"; + errors: { + [key: string]: string; + }; }; /** - * Main Model - Flux2 Klein - * @description Loads a Flux2 Klein model, outputting its submodels. - * - * Flux2 Klein uses Qwen3 as the text encoder instead of CLIP+T5. - * It uses a 32-channel VAE (AutoencoderKLFlux2) instead of the 16-channel FLUX.1 VAE. - * - * When using a Diffusers format model, both VAE and Qwen3 encoder are extracted - * automatically from the main model. You can override with standalone models: - * - Transformer: Always from Flux2 Klein main model - * - VAE: From main model (Diffusers) or standalone VAE - * - Qwen3 Encoder: From main model (Diffusers) or standalone Qwen3 model + * Denoise - SD1.5, SDXL + * @description Denoises noisy latents to decodable images */ - Flux2KleinModelLoaderInvocation: { + DenoiseLatentsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9045,142 +9902,101 @@ export type components = { */ use_cache?: boolean; /** - * Transformer - * @description Flux model (Transformer) to load + * Positive Conditioning + * @description Positive conditioning tensor + * @default null */ - model: components["schemas"]["ModelIdentifierField"]; + positive_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; /** - * VAE - * @description Standalone VAE model. Flux2 Klein uses the same VAE as FLUX (16-channel). If not provided, VAE will be loaded from the Qwen3 Source model. + * Negative Conditioning + * @description Negative conditioning tensor * @default null */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + negative_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; /** - * Qwen3 Encoder - * @description Standalone Qwen3 Encoder model. If not provided, encoder will be loaded from the Qwen3 Source model. + * @description Noise tensor * @default null */ - qwen3_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; + noise?: components["schemas"]["LatentsField"] | null; /** - * Qwen3 Source (Diffusers) - * @description Diffusers Flux2 Klein model to extract VAE and/or Qwen3 encoder from. Use this if you don't have separate VAE/Qwen3 models. Ignored if both VAE and Qwen3 Encoder are provided separately. - * @default null + * Steps + * @description Number of steps to run + * @default 10 */ - qwen3_source_model?: components["schemas"]["ModelIdentifierField"] | null; + steps?: number; /** - * Max Seq Length - * @description Max sequence length for the Qwen3 encoder. - * @default 512 - * @enum {integer} + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 7.5 */ - max_seq_len?: 256 | 512; + cfg_scale?: number | number[]; /** - * type - * @default flux2_klein_model_loader - * @constant + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - type: "flux2_klein_model_loader"; - }; - /** - * Flux2KleinModelLoaderOutput - * @description Flux2 Klein model loader output. - */ - Flux2KleinModelLoaderOutput: { + denoising_start?: number; /** - * Transformer - * @description Transformer + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - transformer: components["schemas"]["TransformerField"]; + denoising_end?: number; /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder + * Scheduler + * @description Scheduler to use during inference + * @default euler + * @enum {string} */ - qwen3_encoder: components["schemas"]["Qwen3EncoderField"]; + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** - * VAE - * @description VAE + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - vae: components["schemas"]["VAEField"]; + unet?: components["schemas"]["UNetField"] | null; /** - * Max Seq Length - * @description The max sequence length for the Qwen3 encoder. - * @enum {integer} + * Control + * @default null */ - max_seq_len: 256 | 512; + control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; /** - * type - * @default flux2_klein_model_loader_output - * @constant + * IP-Adapter + * @description IP-Adapter to apply + * @default null */ - type: "flux2_klein_model_loader_output"; - }; - /** - * Prompt - Flux2 Klein - * @description Encodes and preps a prompt for Flux2 Klein image generation. - * - * Flux2 Klein uses Qwen3 as the text encoder, extracting hidden states from - * layers (9, 18, 27) and stacking them for richer text representations. - * This matches the diffusers Flux2KleinPipeline implementation exactly. - */ - Flux2KleinTextEncoderInvocation: { + ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * T2I-Adapter + * @description T2I-Adapter(s) to apply + * @default null */ - id: string; + t2i_adapter?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * CFG Rescale Multiplier + * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR + * @default 0 */ - is_intermediate?: boolean; + cfg_rescale_multiplier?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Prompt - * @description Text prompt to encode. - * @default null - */ - prompt?: string | null; - /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder + * @description Latents tensor * @default null */ - qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; - /** - * Max Seq Len - * @description Max sequence length for the Qwen3 encoder. - * @default 512 - * @enum {integer} - */ - max_seq_len?: 256 | 512; + latents?: components["schemas"]["LatentsField"] | null; /** - * @description A mask defining the region that this conditioning prompt applies to. + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. * @default null */ - mask?: components["schemas"]["TensorField"] | null; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** * type - * @default flux2_klein_text_encoder + * @default denoise_latents * @constant */ - type: "flux2_klein_text_encoder"; + type: "denoise_latents"; }; - /** - * Latents to Image - FLUX2 - * @description Generates an image from latents using FLUX.2 Klein's 32-channel VAE. - */ - Flux2VaeDecodeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + /** Denoise - SD1.5, SDXL + Metadata */ + DenoiseLatentsMetaInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -9204,119 +10020,151 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * Positive Conditioning + * @description Positive conditioning tensor * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + positive_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; /** - * @description VAE + * Negative Conditioning + * @description Negative conditioning tensor * @default null */ - vae?: components["schemas"]["VAEField"] | null; + negative_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; /** - * type - * @default flux2_vae_decode - * @constant + * @description Noise tensor + * @default null */ - type: "flux2_vae_decode"; - }; - /** - * Image to Latents - FLUX2 - * @description Encodes an image into latents using FLUX.2 Klein's 32-channel VAE. - */ - Flux2VaeEncodeInvocation: { + noise?: components["schemas"]["LatentsField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Steps + * @description Number of steps to run + * @default 10 */ - id: string; + steps?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 7.5 */ - is_intermediate?: boolean; + cfg_scale?: number | number[]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - use_cache?: boolean; + denoising_start?: number; /** - * @description The image to encode. + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 + */ + denoising_end?: number; + /** + * Scheduler + * @description Scheduler to use during inference + * @default euler + * @enum {string} + */ + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + /** + * UNet + * @description UNet (scheduler, LoRAs) * @default null */ - image?: components["schemas"]["ImageField"] | null; + unet?: components["schemas"]["UNetField"] | null; /** - * @description VAE + * Control * @default null */ - vae?: components["schemas"]["VAEField"] | null; + control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; /** - * type - * @default flux2_vae_encode - * @constant + * IP-Adapter + * @description IP-Adapter to apply + * @default null */ - type: "flux2_vae_encode"; - }; - /** - * Flux2VariantType - * @description FLUX.2 model variants. - * @enum {string} - */ - Flux2VariantType: "klein_4b" | "klein_9b" | "klein_9b_base"; - /** - * FluxConditioningCollectionOutput - * @description Base class for nodes that output a collection of conditioning tensors - */ - FluxConditioningCollectionOutput: { + ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * Collection - * @description The output conditioning tensors + * T2I-Adapter + * @description T2I-Adapter(s) to apply + * @default null */ - collection: components["schemas"]["FluxConditioningField"][]; + t2i_adapter?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + /** + * CFG Rescale Multiplier + * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR + * @default 0 + */ + cfg_rescale_multiplier?: number; + /** + * @description Latents tensor + * @default null + */ + latents?: components["schemas"]["LatentsField"] | null; + /** + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null + */ + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** * type - * @default flux_conditioning_collection_output + * @default denoise_latents_meta * @constant */ - type: "flux_conditioning_collection_output"; + type: "denoise_latents_meta"; }; /** - * FluxConditioningField - * @description A conditioning tensor primitive value + * DenoiseMaskField + * @description An inpaint mask field */ - FluxConditioningField: { + DenoiseMaskField: { /** - * Conditioning Name - * @description The name of conditioning tensor + * Mask Name + * @description The name of the mask image */ - conditioning_name: string; + mask_name: string; /** - * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. + * Masked Latents Name + * @description The name of the masked image latents * @default null */ - mask?: components["schemas"]["TensorField"] | null; + masked_latents_name?: string | null; + /** + * Gradient + * @description Used for gradient inpainting + * @default false + */ + gradient?: boolean; }; /** - * FluxConditioningOutput - * @description Base class for nodes that output a single conditioning tensor + * DenoiseMaskOutput + * @description Base class for nodes that output a single image */ - FluxConditioningOutput: { - /** @description Conditioning tensor */ - conditioning: components["schemas"]["FluxConditioningField"]; + DenoiseMaskOutput: { + /** @description Mask for denoise model run */ + denoise_mask: components["schemas"]["DenoiseMaskField"]; /** * type - * @default flux_conditioning_output + * @default denoise_mask_output * @constant */ - type: "flux_conditioning_output"; + type: "denoise_mask_output"; }; /** - * Control LoRA - FLUX - * @description LoRA model and Image to use with FLUX transformer generation. + * Depth Anything Depth Estimation + * @description Generates a depth map using a Depth Anything model. */ - FluxControlLoRALoaderInvocation: { + DepthAnythingDepthEstimationInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9335,368 +10183,336 @@ export type components = { */ use_cache?: boolean; /** - * Control LoRA - * @description Control LoRA model to load - * @default null - */ - lora?: components["schemas"]["ModelIdentifierField"] | null; - /** - * @description The image to encode. + * @description The image to process * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Weight - * @description The weight of the LoRA. - * @default 1 + * Model Size + * @description The size of the depth model to use + * @default small_v2 + * @enum {string} */ - weight?: number; + model_size?: "large" | "base" | "small" | "small_v2"; /** * type - * @default flux_control_lora_loader + * @default depth_anything_depth_estimation * @constant */ - type: "flux_control_lora_loader"; + type: "depth_anything_depth_estimation"; }; /** - * FluxControlLoRALoaderOutput - * @description Flux Control LoRA Loader Output + * Divide Integers + * @description Divides two numbers */ - FluxControlLoRALoaderOutput: { + DivideInvocation: { /** - * Flux Control LoRA - * @description Control LoRAs to apply on model loading - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - control_lora: components["schemas"]["ControlLoRAField"]; + id: string; /** - * type - * @default flux_control_lora_loader_output - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - type: "flux_control_lora_loader_output"; - }; - /** FluxControlNetField */ - FluxControlNetField: { - /** @description The control image */ - image: components["schemas"]["ImageField"]; - /** @description The ControlNet model to use */ - control_model: components["schemas"]["ModelIdentifierField"]; + is_intermediate?: boolean; /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - control_weight?: number | number[]; + use_cache?: boolean; /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) + * A + * @description The first number * @default 0 */ - begin_step_percent?: number; - /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 - */ - end_step_percent?: number; + a?: number; /** - * Resize Mode - * @description The resize mode to use - * @default just_resize - * @enum {string} + * B + * @description The second number + * @default 0 */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + b?: number; /** - * Instantx Control Mode - * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. - * @default -1 + * type + * @default div + * @constant */ - instantx_control_mode?: number | null; + type: "div"; }; /** - * FLUX ControlNet - * @description Collect FLUX ControlNet info to pass to other nodes. + * DownloadCancelledEvent + * @description Event model for download_cancelled */ - FluxControlNetInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + DownloadCancelledEvent: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Timestamp + * @description The timestamp of the event */ - use_cache?: boolean; + timestamp: number; /** - * @description The control image - * @default null + * Source + * @description The source of the download */ - image?: components["schemas"]["ImageField"] | null; + source: string; + }; + /** + * DownloadCompleteEvent + * @description Event model for download_complete + */ + DownloadCompleteEvent: { /** - * @description ControlNet model to load - * @default null + * Timestamp + * @description The timestamp of the event */ - control_model?: components["schemas"]["ModelIdentifierField"] | null; + timestamp: number; /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * Source + * @description The source of the download */ - control_weight?: number | number[]; + source: string; /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 + * Download Path + * @description The local path where the download is saved */ - begin_step_percent?: number; + download_path: string; /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Total Bytes + * @description The total number of bytes downloaded */ - end_step_percent?: number; + total_bytes: number; + }; + /** + * DownloadErrorEvent + * @description Event model for download_error + */ + DownloadErrorEvent: { /** - * Resize Mode - * @description The resize mode used - * @default just_resize - * @enum {string} + * Timestamp + * @description The timestamp of the event */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + timestamp: number; /** - * Instantx Control Mode - * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. - * @default -1 + * Source + * @description The source of the download */ - instantx_control_mode?: number | null; + source: string; /** - * type - * @default flux_controlnet - * @constant + * Error Type + * @description The type of error */ - type: "flux_controlnet"; - }; - /** - * FluxControlNetOutput - * @description FLUX ControlNet info - */ - FluxControlNetOutput: { - /** @description ControlNet(s) to apply */ - control: components["schemas"]["FluxControlNetField"]; + error_type: string; /** - * type - * @default flux_controlnet_output - * @constant + * Error + * @description The error message */ - type: "flux_controlnet_output"; + error: string; }; /** - * FLUX Denoise - * @description Run denoising process with a FLUX transformer model. + * DownloadJob + * @description Class to monitor and control a model download request. */ - FluxDenoiseInvocation: { + DownloadJob: { /** * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description Numeric ID of this job + * @default -1 */ - id: string; + id?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Dest + * Format: path + * @description Initial destination of downloaded model on local disk; a directory or file path */ - is_intermediate?: boolean; + dest: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Download Path + * @description Final location of downloaded file or directory */ - use_cache?: boolean; + download_path?: string | null; /** - * @description Latents tensor - * @default null + * @description Status of the download + * @default waiting */ - latents?: components["schemas"]["LatentsField"] | null; + status?: components["schemas"]["DownloadJobStatus"]; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. - * @default null + * Bytes + * @description Bytes downloaded so far + * @default 0 */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + bytes?: number; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps + * Total Bytes + * @description Total file size (bytes) * @default 0 */ - denoising_start?: number; + total_bytes?: number; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * Error Type + * @description Name of exception that caused an error */ - denoising_end?: number; + error_type?: string | null; /** - * Add Noise - * @description Add noise based on denoising start. - * @default true + * Error + * @description Traceback of the exception that caused an error */ - add_noise?: boolean; + error?: string | null; /** - * Transformer - * @description Flux model (Transformer) to load - * @default null + * Source + * Format: uri + * @description Where to download from. Specific types specified in child classes. */ - transformer?: components["schemas"]["TransformerField"] | null; + source: string; /** - * Control LoRA - * @description Control LoRA model to load - * @default null + * Access Token + * @description authorization token for protected resources */ - control_lora?: components["schemas"]["ControlLoRAField"] | null; + access_token?: string | null; /** - * Positive Text Conditioning - * @description Positive conditioning tensor - * @default null + * Priority + * @description Queue priority; lower values are higher priority + * @default 10 */ - positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + priority?: number; /** - * Negative Text Conditioning - * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. - * @default null + * Job Started + * @description Timestamp for when the download job started */ - negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + job_started?: string | null; /** - * Redux Conditioning - * @description FLUX Redux conditioning tensor. - * @default null + * Job Ended + * @description Timestamp for when the download job ende1d (completed or errored) */ - redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; + job_ended?: string | null; /** - * @description FLUX Fill conditioning. - * @default null + * Content Type + * @description Content type of downloaded file */ - fill_conditioning?: components["schemas"]["FluxFillConditioningField"] | null; + content_type?: string | null; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 1 + * Canonical Url + * @description Canonical URL to request on resume */ - cfg_scale?: number | number[]; + canonical_url?: string | null; /** - * CFG Scale Start Step - * @description Index of the first step to apply cfg_scale. Negative indices count backwards from the the last step (e.g. a value of -1 refers to the final step). - * @default 0 + * Etag + * @description ETag from the remote server, if available */ - cfg_scale_start_step?: number; + etag?: string | null; /** - * CFG Scale End Step - * @description Index of the last step to apply cfg_scale. Negative indices count backwards from the last step (e.g. a value of -1 refers to the final step). - * @default -1 + * Last Modified + * @description Last-Modified from the remote server, if available */ - cfg_scale_end_step?: number; + last_modified?: string | null; /** - * Width - * @description Width of the generated image. - * @default 1024 + * Final Url + * @description Final resolved URL after redirects, if available */ - width?: number; + final_url?: string | null; /** - * Height - * @description Height of the generated image. - * @default 1024 + * Expected Total Bytes + * @description Expected total size of the download */ - height?: number; + expected_total_bytes?: number | null; /** - * Num Steps - * @description Number of diffusion steps. Recommended values are schnell: 4, dev: 50. - * @default 4 + * Resume Required + * @description True if server refused resume; restart required + * @default false */ - num_steps?: number; + resume_required?: boolean; /** - * Scheduler - * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. - * @default euler - * @enum {string} + * Resume Message + * @description Message explaining why resume is required */ - scheduler?: "euler" | "heun" | "lcm"; + resume_message?: string | null; /** - * Guidance - * @description The guidance strength. Higher values adhere more strictly to the prompt, and will produce less diverse images. FLUX dev only, ignored for schnell. - * @default 4 + * Resume From Scratch + * @description True if resume metadata existed but the partial file was missing and the download restarted from the beginning + * @default false */ - guidance?: number; + resume_from_scratch?: boolean; + }; + /** + * DownloadJobStatus + * @description State of a download job. + * @enum {string} + */ + DownloadJobStatus: "waiting" | "running" | "paused" | "completed" | "cancelled" | "error"; + /** + * DownloadPausedEvent + * @description Event model for download_paused + */ + DownloadPausedEvent: { /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * Timestamp + * @description The timestamp of the event */ - seed?: number; + timestamp: number; /** - * Control - * @description ControlNet models. - * @default null + * Source + * @description The source of the download */ - control?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; + source: string; + }; + /** + * DownloadProgressEvent + * @description Event model for download_progress + */ + DownloadProgressEvent: { /** - * @description VAE - * @default null + * Timestamp + * @description The timestamp of the event */ - controlnet_vae?: components["schemas"]["VAEField"] | null; + timestamp: number; /** - * IP-Adapter - * @description IP-Adapter to apply - * @default null + * Source + * @description The source of the download */ - ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + source: string; /** - * Kontext Conditioning - * @description FLUX Kontext conditioning (reference image). - * @default null + * Download Path + * @description The local path where the download is saved */ - kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; + download_path: string; /** - * Dype Preset - * @description DyPE preset for high-resolution generation. 'auto' enables automatically for resolutions > 1536px. 'area' enables automatically based on image area. '4k' uses optimized settings for 4K output. - * @default off - * @enum {string} + * Current Bytes + * @description The number of bytes downloaded so far */ - dype_preset?: "off" | "manual" | "auto" | "area" | "4k"; + current_bytes: number; /** - * Dype Scale - * @description DyPE magnitude (λs). Higher values = stronger extrapolation. Only used when dype_preset is not 'off'. - * @default null + * Total Bytes + * @description The total number of bytes to be downloaded */ - dype_scale?: number | null; + total_bytes: number; + }; + /** + * DownloadStartedEvent + * @description Event model for download_started + */ + DownloadStartedEvent: { /** - * Dype Exponent - * @description DyPE decay speed (λt). Controls transition from low to high frequency detail. Only used when dype_preset is not 'off'. - * @default null + * Timestamp + * @description The timestamp of the event */ - dype_exponent?: number | null; + timestamp: number; /** - * type - * @default flux_denoise - * @constant + * Source + * @description The source of the download */ - type: "flux_denoise"; + source: string; + /** + * Download Path + * @description The local path where the download is saved + */ + download_path: string; }; /** - * FLUX Denoise + Metadata - * @description Run denoising process with a FLUX transformer model + metadata. + * Dynamic Prompt + * @description Parses a prompt using adieyal/dynamicprompts' random or combinatorial generator */ - FluxDenoiseLatentsMetaInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + DynamicPromptInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9711,191 +10527,238 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * Prompt + * @description The prompt to parse with dynamicprompts * @default null */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + prompt?: string | null; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Max Prompts + * @description The number of prompts to generate + * @default 1 */ - denoising_start?: number; + max_prompts?: number; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * Combinatorial + * @description Whether to use the combinatorial generator + * @default false */ - denoising_end?: number; + combinatorial?: boolean; /** - * Add Noise - * @description Add noise based on denoising start. - * @default true + * type + * @default dynamic_prompt + * @constant */ - add_noise?: boolean; + type: "dynamic_prompt"; + }; + /** DynamicPromptsResponse */ + DynamicPromptsResponse: { + /** Prompts */ + prompts: string[]; + /** Error */ + error?: string | null; + }; + /** + * Upscale (RealESRGAN) + * @description Upscales an image using RealESRGAN. + */ + ESRGANInvocation: { /** - * Transformer - * @description Flux model (Transformer) to load + * @description The board to save the image to * @default null */ - transformer?: components["schemas"]["TransformerField"] | null; + board?: components["schemas"]["BoardField"] | null; /** - * Control LoRA - * @description Control LoRA model to load + * @description Optional metadata to be saved with the image * @default null */ - control_lora?: components["schemas"]["ControlLoRAField"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Positive Text Conditioning - * @description Positive conditioning tensor - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + id: string; /** - * Negative Text Conditioning - * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + is_intermediate?: boolean; /** - * Redux Conditioning - * @description FLUX Redux conditioning tensor. - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; + use_cache?: boolean; /** - * @description FLUX Fill conditioning. + * @description The input image * @default null */ - fill_conditioning?: components["schemas"]["FluxFillConditioningField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 1 + * Model Name + * @description The Real-ESRGAN model to use + * @default RealESRGAN_x4plus.pth + * @enum {string} */ - cfg_scale?: number | number[]; + model_name?: "RealESRGAN_x4plus.pth" | "RealESRGAN_x4plus_anime_6B.pth" | "ESRGAN_SRx4_DF2KOST_official-ff704c30.pth" | "RealESRGAN_x2plus.pth"; /** - * CFG Scale Start Step - * @description Index of the first step to apply cfg_scale. Negative indices count backwards from the the last step (e.g. a value of -1 refers to the final step). - * @default 0 + * Tile Size + * @description Tile size for tiled ESRGAN upscaling (0=tiling disabled) + * @default 400 */ - cfg_scale_start_step?: number; - /** - * CFG Scale End Step - * @description Index of the last step to apply cfg_scale. Negative indices count backwards from the last step (e.g. a value of -1 refers to the final step). - * @default -1 - */ - cfg_scale_end_step?: number; + tile_size?: number; /** - * Width - * @description Width of the generated image. - * @default 1024 + * type + * @default esrgan + * @constant */ - width?: number; + type: "esrgan"; + }; + /** Edge */ + Edge: { + /** @description The connection for the edge's from node and field */ + source: components["schemas"]["EdgeConnection"]; + /** @description The connection for the edge's to node and field */ + destination: components["schemas"]["EdgeConnection"]; + }; + /** EdgeConnection */ + EdgeConnection: { /** - * Height - * @description Height of the generated image. - * @default 1024 + * Node Id + * @description The id of the node for this edge connection */ - height?: number; + node_id: string; /** - * Num Steps - * @description Number of diffusion steps. Recommended values are schnell: 4, dev: 50. - * @default 4 + * Field + * @description The field for this connection */ - num_steps?: number; + field: string; + }; + /** + * Enhance Detail + * @description Enhance Detail using guided filter + */ + EnhanceDetailInvocation: { /** - * Scheduler - * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. - * @default euler - * @enum {string} + * @description The board to save the image to + * @default null */ - scheduler?: "euler" | "heun" | "lcm"; + board?: components["schemas"]["BoardField"] | null; /** - * Guidance - * @description The guidance strength. Higher values adhere more strictly to the prompt, and will produce less diverse images. FLUX dev only, ignored for schnell. - * @default 4 + * @description Optional metadata to be saved with the image + * @default null */ - guidance?: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - seed?: number; + id: string; /** - * Control - * @description ControlNet models. - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - control?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; + is_intermediate?: boolean; /** - * @description VAE - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - controlnet_vae?: components["schemas"]["VAEField"] | null; + use_cache?: boolean; /** - * IP-Adapter - * @description IP-Adapter to apply + * @description The image to detail enhance * @default null */ - ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Kontext Conditioning - * @description FLUX Kontext conditioning (reference image). - * @default null + * Filter Radius + * @description radius of filter + * @default 2 */ - kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; + filter_radius?: number; /** - * Dype Preset - * @description DyPE preset for high-resolution generation. 'auto' enables automatically for resolutions > 1536px. 'area' enables automatically based on image area. '4k' uses optimized settings for 4K output. - * @default off - * @enum {string} + * Sigma + * @description sigma + * @default 0.1 */ - dype_preset?: "off" | "manual" | "auto" | "area" | "4k"; + sigma?: number; /** - * Dype Scale - * @description DyPE magnitude (λs). Higher values = stronger extrapolation. Only used when dype_preset is not 'off'. - * @default null + * Denoise + * @description denoise + * @default 0.1 */ - dype_scale?: number | null; + denoise?: number; /** - * Dype Exponent - * @description DyPE decay speed (λt). Controls transition from low to high frequency detail. Only used when dype_preset is not 'off'. - * @default null + * Detail Multiplier + * @description detail multiplier + * @default 2 */ - dype_exponent?: number | null; + detail_multiplier?: number; /** * type - * @default flux_denoise_meta + * @default enhance_detail * @constant */ - type: "flux_denoise_meta"; + type: "enhance_detail"; + }; + /** EnqueueBatchResult */ + EnqueueBatchResult: { + /** + * Queue Id + * @description The ID of the queue + */ + queue_id: string; + /** + * Enqueued + * @description The total number of queue items enqueued + */ + enqueued: number; + /** + * Requested + * @description The total number of queue items requested to be enqueued + */ + requested: number; + /** @description The batch that was enqueued */ + batch: components["schemas"]["Batch"]; + /** + * Priority + * @description The priority of the enqueued batch + */ + priority: number; + /** + * Item Ids + * @description The IDs of the queue items that were enqueued + */ + item_ids: number[]; }; /** - * FluxFillConditioningField - * @description A FLUX Fill conditioning field. + * EscapedOutput + * @description The input str with double quotes escaped */ - FluxFillConditioningField: { - /** @description The FLUX Fill reference image. */ - image: components["schemas"]["ImageField"]; - /** @description The FLUX Fill inpaint mask. */ - mask: components["schemas"]["TensorField"]; + EscapedOutput: { + /** + * Prompt + * @description The input str with double quotes escaped + */ + prompt: string; + /** + * type + * @default escaped_quotes_output + * @constant + */ + type: "escaped_quotes_output"; }; /** - * FLUX Fill Conditioning - * @description Prepare the FLUX Fill conditioning data. + * Quote Escaper + * @description Escapes double quotes from input strings */ - FluxFillInvocation: { + EscaperInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9910,48 +10773,27 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * @description The FLUX Fill reference image. - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The bool inpainting mask. Excluded regions should be set to False, included regions should be set to True. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default flux_fill - * @constant - */ - type: "flux_fill"; - }; - /** - * FluxFillOutput - * @description The conditioning output of a FLUX Fill invocation. - */ - FluxFillOutput: { - /** - * Conditioning - * @description FLUX Redux conditioning tensor + * Prompt + * @description the string to escape quotes from + * @default */ - fill_cond: components["schemas"]["FluxFillConditioningField"]; + prompt?: string; /** * type - * @default flux_fill_output + * @default quote_escaper * @constant */ - type: "flux_fill_output"; + type: "quote_escaper"; }; /** - * FLUX IP-Adapter - * @description Collects FLUX IP-Adapter info to pass to other nodes. + * Even Split XYImage Tile Generator + * @description Cuts up an image into a number of even sized tiles with the overlap been a percentage of the tile size and outputs a string representation of the tiles to use */ - FluxIPAdapterInvocation: { + EvenSplitXYTileGenerator: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9970,54 +10812,65 @@ export type components = { */ use_cache?: boolean; /** - * @description The IP-Adapter image prompt(s). + * @description The input image * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * IP-Adapter Model - * @description The IP-Adapter model. - * @default null + * Num X Tiles + * @description Number of tiles to divide image into on the x axis + * @default 2 */ - ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; + num_x_tiles?: number; /** - * Clip Vision Model - * @description CLIP Vision model to use. - * @default ViT-L - * @constant + * Num Y Tiles + * @description Number of tiles to divide image into on the y axis + * @default 2 */ - clip_vision_model?: "ViT-L"; + num_y_tiles?: number; /** - * Weight - * @description The weight given to the IP-Adapter - * @default 1 + * Overlap + * @description Overlap amount of tile size (0-1) + * @default 0.25 */ - weight?: number | number[]; + overlap?: number; /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) - * @default 0 + * type + * @default even_split_xy_tile_generator + * @constant */ - begin_step_percent?: number; + type: "even_split_xy_tile_generator"; + }; + /** + * EvolutionaryPromptListOutput + * @description Output for a list of generated prompts and their seed vectors. + */ + EvolutionaryPromptListOutput: { /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) - * @default 1 + * Prompt Seed Pairs + * @description List of JSON strings, each containing [prompt, seed_vector_json] */ - end_step_percent?: number; + prompt_seed_pairs: string[]; + /** + * Selected Pair + * @description The index-selected prompt from the output population + */ + selected_pair: string; /** * type - * @default flux_ip_adapter + * @default evolutionary_prompt_list_output * @constant */ - type: "flux_ip_adapter"; + type: "evolutionary_prompt_list_output"; }; /** - * FLUX Kontext Image Prep - * @description Prepares an image or images for use with FLUX Kontext. The first/single image is resized to the nearest - * preferred Kontext resolution. All other images are concatenated horizontally, maintaining their aspect ratio. + * Expand Mask with Fade + * @description Expands a mask with a fade effect. The mask uses black to indicate areas to keep from the generated image and white for areas to discard. + * The mask is thresholded to create a binary mask, and then a distance transform is applied to create a fade effect. + * The fade size is specified in pixels, and the mask is expanded by that amount. The result is a mask with a smooth transition from black to white. + * If the fade size is 0, the mask is returned as-is. */ - FluxKontextConcatenateImagesInvocation: { + ExpandMaskWithFadeInvocation: { /** * @description The board to save the image to * @default null @@ -10046,37 +10899,42 @@ export type components = { */ use_cache?: boolean; /** - * Images - * @description The images to concatenate + * @description The mask to expand * @default null */ - images?: components["schemas"]["ImageField"][] | null; + mask?: components["schemas"]["ImageField"] | null; /** - * Use Preferred Resolution - * @description Use FLUX preferred resolutions for the first image - * @default true + * Threshold + * @description The threshold for the binary mask (0-255) + * @default 0 */ - use_preferred_resolution?: boolean; + threshold?: number; /** - * type - * @default flux_kontext_image_prep - * @constant - */ - type: "flux_kontext_image_prep"; + * Fade Size Px + * @description The size of the fade in pixels + * @default 32 + */ + fade_size_px?: number; + /** + * type + * @default expand_mask_with_fade + * @constant + */ + type: "expand_mask_with_fade"; }; - /** - * FluxKontextConditioningField - * @description A conditioning field for FLUX Kontext (reference image). - */ - FluxKontextConditioningField: { - /** @description The Kontext reference image. */ - image: components["schemas"]["ImageField"]; + /** ExposedField */ + ExposedField: { + /** Nodeid */ + nodeId: string; + /** Fieldname */ + fieldName: string; }; /** - * Kontext Conditioning - FLUX - * @description Prepares a reference image for FLUX Kontext conditioning. + * Extract Image Collection Metadata (Bool) + * @description This node extracts specified metadata values as booleans from a collection of input images. + * Values are converted to boolean: truthy values become True, falsy values (including None, empty string, 0) become False. */ - FluxKontextInvocation: { + ExtractImageCollectionMetadataBooleanInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10095,39 +10953,30 @@ export type components = { */ use_cache?: boolean; /** - * @description The Kontext reference image. + * Image Collection + * @description A collection of images from which to extract metadata. * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default flux_kontext - * @constant - */ - type: "flux_kontext"; - }; - /** - * FluxKontextOutput - * @description The conditioning output of a FLUX Kontext invocation. - */ - FluxKontextOutput: { + images?: components["schemas"]["ImageField"][] | null; /** - * Kontext Conditioning - * @description FLUX Kontext conditioning (reference image) + * Metadata Key + * @description Metadata key to extract values for Output. Leave empty to ignore. + * @default */ - kontext_cond: components["schemas"]["FluxKontextConditioningField"]; + key?: string; /** * type - * @default flux_kontext_output + * @default extract_image_collection_metadata_boolean * @constant */ - type: "flux_kontext_output"; + type: "extract_image_collection_metadata_boolean"; }; /** - * Apply LoRA - FLUX - * @description Apply a LoRA model to a FLUX transformer and/or text encoder. + * Extract Image Collection Metadata (Float) + * @description This node extracts specified metadata values as floats from a collection of input images. + * Non-float values will attempt to be converted. If conversion fails, 0.0 is used. */ - FluxLoRALoaderInvocation: { + ExtractImageCollectionMetadataFloatInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10146,77 +10995,30 @@ export type components = { */ use_cache?: boolean; /** - * LoRA - * @description LoRA model to load - * @default null - */ - lora?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 - */ - weight?: number; - /** - * FLUX Transformer - * @description Transformer - * @default null - */ - transformer?: components["schemas"]["TransformerField"] | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip?: components["schemas"]["CLIPField"] | null; - /** - * T5 Encoder - * @description T5 tokenizer and text encoder - * @default null - */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; - /** - * type - * @default flux_lora_loader - * @constant - */ - type: "flux_lora_loader"; - }; - /** - * FluxLoRALoaderOutput - * @description FLUX LoRA Loader Output - */ - FluxLoRALoaderOutput: { - /** - * FLUX Transformer - * @description Transformer - * @default null - */ - transformer: components["schemas"]["TransformerField"] | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Image Collection + * @description A collection of images from which to extract metadata. * @default null */ - clip: components["schemas"]["CLIPField"] | null; + images?: components["schemas"]["ImageField"][] | null; /** - * T5 Encoder - * @description T5 tokenizer and text encoder - * @default null + * Metadata Key + * @description Metadata key to extract values for Output. Leave empty to ignore. + * @default */ - t5_encoder: components["schemas"]["T5EncoderField"] | null; + key?: string; /** * type - * @default flux_lora_loader_output + * @default extract_image_collection_metadata_float * @constant */ - type: "flux_lora_loader_output"; + type: "extract_image_collection_metadata_float"; }; /** - * Main Model - FLUX - * @description Loads a flux base model, outputting its submodels. + * Extract Image Collection Metadata (Int) + * @description This node extracts specified metadata values as integers from a collection of input images. + * Non-integer values will attempt to be converted. If conversion fails, 0 is used. */ - FluxModelLoaderInvocation: { + ExtractImageCollectionMetadataIntegerInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10235,91 +11037,70 @@ export type components = { */ use_cache?: boolean; /** - * @description Flux model (Transformer) to load - * @default null - */ - model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * T5 Encoder - * @description T5 tokenizer and text encoder - * @default null - */ - t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * CLIP Embed - * @description CLIP Embed loader + * Image Collection + * @description A collection of images from which to extract metadata. * @default null */ - clip_embed_model?: components["schemas"]["ModelIdentifierField"] | null; + images?: components["schemas"]["ImageField"][] | null; /** - * VAE - * @description VAE model to load - * @default null + * Metadata Key + * @description Metadata key to extract values for Output. Leave empty to ignore. + * @default */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + key?: string; /** * type - * @default flux_model_loader + * @default extract_image_collection_metadata_integer * @constant */ - type: "flux_model_loader"; + type: "extract_image_collection_metadata_integer"; }; /** - * FluxModelLoaderOutput - * @description Flux base model loader output + * Extract Image Collection Metadata (String) + * @description This node extracts specified metadata values as strings from a collection of input images. */ - FluxModelLoaderOutput: { + ExtractImageCollectionMetadataStringInvocation: { /** - * Transformer - * @description Transformer + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - transformer: components["schemas"]["TransformerField"]; + id: string; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - clip: components["schemas"]["CLIPField"]; + is_intermediate?: boolean; /** - * T5 Encoder - * @description T5 tokenizer and text encoder + * Use Cache + * @description Whether or not to use the cache + * @default true */ - t5_encoder: components["schemas"]["T5EncoderField"]; + use_cache?: boolean; /** - * VAE - * @description VAE + * Image Collection + * @description A collection of images from which to extract metadata. + * @default null */ - vae: components["schemas"]["VAEField"]; + images?: components["schemas"]["ImageField"][] | null; /** - * Max Seq Length - * @description The max sequence length to used for the T5 encoder. (256 for schnell transformer, 512 for dev transformer) - * @enum {integer} + * Metadata Key + * @description Metadata key to extract values for Output. Leave empty to ignore. + * @default */ - max_seq_len: 256 | 512; + key?: string; /** * type - * @default flux_model_loader_output + * @default extract_image_collection_metadata_string * @constant */ - type: "flux_model_loader_output"; - }; - /** - * FluxReduxConditioningField - * @description A FLUX Redux conditioning tensor primitive value - */ - FluxReduxConditioningField: { - /** @description The Redux image conditioning tensor. */ - conditioning: components["schemas"]["TensorField"]; - /** - * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; + type: "extract_image_collection_metadata_string"; }; /** - * FLUX Redux - * @description Runs a FLUX Redux model to generate a conditioning tensor. + * Extrude Depth from Mask + * @description Node for creating fake depth by "extruding" a mask using OpenCV. */ - FluxReduxInvocation: { + ExtrudeDepthInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10334,73 +11115,74 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * @description The FLUX Redux image prompt. + * @description The mask from which to extrude * @default null */ - image?: components["schemas"]["ImageField"] | null; + mask?: components["schemas"]["ImageField"] | null; /** - * @description The bool mask associated with this FLUX Redux image prompt. Excluded regions should be set to False, included regions should be set to True. - * @default null + * Direction + * @description Extrude direction in degrees + * @default 45 */ - mask?: components["schemas"]["TensorField"] | null; + direction?: number; /** - * FLUX Redux Model - * @description The FLUX Redux model to use. - * @default null + * Shift + * @description Number of pixels to shift bottom from top + * @default 40 */ - redux_model?: components["schemas"]["ModelIdentifierField"] | null; + shift?: number; /** - * Downsampling Factor - * @description Redux Downsampling Factor (1-9) - * @default 1 + * Close Point + * @description Closest extrusion depth + * @default 180 */ - downsampling_factor?: number; + close_point?: number; /** - * Downsampling Function - * @description Redux Downsampling Function - * @default area - * @enum {string} + * Far Point + * @description Farthest extrusion depth + * @default 80 */ - downsampling_function?: "nearest" | "bilinear" | "bicubic" | "area" | "nearest-exact"; + far_point?: number; /** - * Weight - * @description Redux weight (0.0-1.0) - * @default 1 + * Bg Threshold + * @description Background threshold + * @default 10 */ - weight?: number; + bg_threshold?: number; /** - * type - * @default flux_redux - * @constant + * Bg Depth + * @description Target background depth + * @default 0 */ - type: "flux_redux"; - }; - /** - * FluxReduxOutput - * @description The conditioning output of a FLUX Redux invocation. - */ - FluxReduxOutput: { + bg_depth?: number; /** - * Conditioning - * @description FLUX Redux conditioning tensor + * Steps + * @description Number of steps in extrusion gradient + * @default 100 */ - redux_cond: components["schemas"]["FluxReduxConditioningField"]; + steps?: number; + /** + * Invert + * @description Inverts mask image before extruding + * @default false + */ + invert?: boolean; /** * type - * @default flux_redux_output + * @default cv_extrude_depth * @constant */ - type: "flux_redux_output"; + type: "cv_extrude_depth"; }; /** - * Prompt - FLUX - * @description Encodes and preps a prompt for a flux image. + * FLUX Conditioning Collection Toggle + * @description Allows boolean selection between two separate FLUX conditioning collection inputs */ - FluxTextEncoderInvocation: { + FLUXConditioningCollectionToggleInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10419,56 +11201,35 @@ export type components = { */ use_cache?: boolean; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Use Second + * @description Use 2nd Input + * @default false */ - clip?: components["schemas"]["CLIPField"] | null; + use_second?: boolean; /** - * T5Encoder - * @description T5 tokenizer and text encoder + * Col1 + * @description First FLUX Conditioning Collection Input * @default null */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; + col1?: components["schemas"]["FluxConditioningField"][] | null; /** - * T5 Max Seq Len - * @description Max sequence length for the T5 encoder. Expected to be 256 for FLUX schnell models and 512 for FLUX dev models. + * Col2 + * @description Second FLUX Conditioning Collection Input * @default null */ - t5_max_seq_len?: (256 | 512) | null; + col2?: components["schemas"]["FluxConditioningField"][] | null; /** - * Prompt - * @description Text prompt to encode. - * @default null + * type + * @default flux_conditioning_collection_toggle + * @constant */ - prompt?: string | null; - /** - * @description A mask defining the region that this conditioning prompt applies to. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default flux_text_encoder - * @constant - */ - type: "flux_text_encoder"; + type: "flux_conditioning_collection_toggle"; }; /** - * Latents to Image - FLUX - * @description Generates an image from latents. + * FLUX Conditioning Toggle + * @description Allows boolean selection between two separate FLUX conditioning inputs */ - FluxVaeDecodeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FLUXConditioningToggleInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10487,27 +11248,33 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * @description First FLUX Conditioning Input * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + cond1?: components["schemas"]["FluxConditioningField"] | null; /** - * @description VAE + * @description Second FLUX Conditioning Input * @default null */ - vae?: components["schemas"]["VAEField"] | null; + cond2?: components["schemas"]["FluxConditioningField"] | null; /** * type - * @default flux_vae_decode + * @default flux_conditioning_toggle * @constant */ - type: "flux_vae_decode"; + type: "flux_conditioning_toggle"; }; /** - * Image to Latents - FLUX - * @description Encodes an image into latents. + * Apply LoRA Collection - FLUX + * @description Applies a collection of LoRAs to a FLUX transformer. */ - FluxVaeEncodeInvocation: { + FLUXLoRACollectionLoader: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10526,78 +11293,122 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to encode. + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. * @default null */ - image?: components["schemas"]["ImageField"] | null; + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * @description VAE + * Transformer + * @description Transformer * @default null */ - vae?: components["schemas"]["VAEField"] | null; + transformer?: components["schemas"]["TransformerField"] | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * T5 Encoder + * @description T5 tokenizer and text encoder + * @default null + */ + t5_encoder?: components["schemas"]["T5EncoderField"] | null; /** * type - * @default flux_vae_encode + * @default flux_lora_collection_loader * @constant */ - type: "flux_vae_encode"; + type: "flux_lora_collection_loader"; }; /** - * FluxVariantType - * @description FLUX.1 model variants. - * @enum {string} + * FLUXRedux_Checkpoint_Config + * @description Model config for FLUX Tools Redux model. */ - FluxVariantType: "schnell" | "dev" | "dev_fill"; - /** FoundModel */ - FoundModel: { + FLUXRedux_Checkpoint_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; /** * Path - * @description Path to the model + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ path: string; /** - * Is Installed - * @description Whether or not the model is already installed + * File Size + * @description The size of the model in bytes. */ - is_installed: boolean; - }; - /** - * FreeUConfig - * @description Configuration for the FreeU hyperparameters. - * - https://huggingface.co/docs/diffusers/main/en/using-diffusers/freeu - * - https://github.com/ChenyangSi/FreeU - */ - FreeUConfig: { + file_size: number; /** - * S1 - * @description Scaling factor for stage 1 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. + * Name + * @description Name of the model. */ - s1: number; + name: string; /** - * S2 - * @description Scaling factor for stage 2 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. + * Description + * @description Model description */ - s2: number; + description: string | null; /** - * B1 - * @description Scaling factor for stage 1 to amplify the contributions of backbone features. + * Source + * @description The original source of the model (path, URL or repo_id). */ - b1: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * B2 - * @description Scaling factor for stage 2 to amplify the contributions of backbone features. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - b2: number; + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default flux_redux + * @constant + */ + type: "flux_redux"; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; }; /** - * Apply FreeU - SD1.5, SDXL - * @description Applies FreeU to the UNet. Suggested values (b1/b2/s1/s2): - * - * SD1.5: 1.2/1.4/0.9/0.2, - * SD2: 1.1/1.2/0.9/0.2, - * SDXL: 1.1/1.2/0.6/0.4, + * FaceIdentifier + * @description Outputs an image with detected face IDs printed on each face. For use with other FaceTools. */ - FreeUInvocation: { + FaceIdentifierInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10616,58 +11427,39 @@ export type components = { */ use_cache?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) + * @description Image to face detect * @default null */ - unet?: components["schemas"]["UNetField"] | null; - /** - * B1 - * @description Scaling factor for stage 1 to amplify the contributions of backbone features. - * @default 1.2 - */ - b1?: number; - /** - * B2 - * @description Scaling factor for stage 2 to amplify the contributions of backbone features. - * @default 1.4 - */ - b2?: number; + image?: components["schemas"]["ImageField"] | null; /** - * S1 - * @description Scaling factor for stage 1 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. - * @default 0.9 + * Minimum Confidence + * @description Minimum confidence for face detection (lower if detection is failing) + * @default 0.5 */ - s1?: number; + minimum_confidence?: number; /** - * S2 - * @description Scaling factor for stage 2 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. - * @default 0.2 + * Chunk + * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. + * @default false */ - s2?: number; + chunk?: boolean; /** * type - * @default freeu + * @default face_identifier * @constant */ - type: "freeu"; + type: "face_identifier"; }; /** - * GeneratePasswordResponse - * @description Response containing a generated password. + * FaceMask + * @description Face mask creation using mediapipe face detection */ - GeneratePasswordResponse: { + FaceMaskInvocation: { /** - * Password - * @description Generated strong password + * @description Optional metadata to be saved with the image + * @default null */ - password: string; - }; - /** - * Get Image Mask Bounding Box - * @description Gets the bounding box of the given mask image. - */ - GetMaskBoundingBoxInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10686,143 +11478,89 @@ export type components = { */ use_cache?: boolean; /** - * @description The mask to crop. + * @description Image to face detect * @default null */ - mask?: components["schemas"]["ImageField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Margin - * @description Margin to add to the bounding box. - * @default 0 + * Face Ids + * @description Comma-separated list of face ids to mask eg '0,2,7'. Numbered from 0. Leave empty to mask all. Find face IDs with FaceIdentifier node. + * @default */ - margin?: number; + face_ids?: string; /** - * @description Color of the mask in the image. - * @default { - * "r": 255, - * "g": 255, - * "b": 255, - * "a": 255 - * } + * Minimum Confidence + * @description Minimum confidence for face detection (lower if detection is failing) + * @default 0.5 */ - mask_color?: components["schemas"]["ColorField"]; + minimum_confidence?: number; /** - * type - * @default get_image_mask_bounding_box - * @constant + * X Offset + * @description Offset for the X-axis of the face mask + * @default 0 */ - type: "get_image_mask_bounding_box"; - }; - /** GlmEncoderField */ - GlmEncoderField: { - /** @description Info to load tokenizer submodel */ - tokenizer: components["schemas"]["ModelIdentifierField"]; - /** @description Info to load text_encoder submodel */ - text_encoder: components["schemas"]["ModelIdentifierField"]; - }; - /** - * GradientMaskOutput - * @description Outputs a denoise mask and an image representing the total gradient of the mask. - */ - GradientMaskOutput: { - /** @description Mask for denoise model run. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. */ - denoise_mask: components["schemas"]["DenoiseMaskField"]; - /** @description Image representing the total gradient area of the mask. For paste-back purposes. */ - expanded_mask_area: components["schemas"]["ImageField"]; + x_offset?: number; /** - * type - * @default gradient_mask_output - * @constant + * Y Offset + * @description Offset for the Y-axis of the face mask + * @default 0 */ - type: "gradient_mask_output"; - }; - /** Graph */ - Graph: { + y_offset?: number; /** - * Id - * @description The id of this graph + * Chunk + * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. + * @default false */ - id?: string; + chunk?: boolean; /** - * Nodes - * @description The nodes in this graph + * Invert Mask + * @description Toggle to invert the mask + * @default false */ - nodes?: { - [key: string]: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; - }; + invert_mask?: boolean; /** - * Edges - * @description The connections between nodes and their fields in this graph + * type + * @default face_mask_detection + * @constant */ - edges?: components["schemas"]["Edge"][]; + type: "face_mask_detection"; }; /** - * GraphExecutionState - * @description Tracks the state of a graph execution + * FaceMaskOutput + * @description Base class for FaceMask output */ - GraphExecutionState: { - /** - * Id - * @description The id of the execution state - */ - id: string; - /** @description The graph being executed */ - graph: components["schemas"]["Graph"]; - /** @description The expanded graph of activated and executed nodes */ - execution_graph: components["schemas"]["Graph"]; - /** - * Executed - * @description The set of node ids that have been executed - */ - executed: string[]; - /** - * Executed History - * @description The list of node ids that have been executed, in order of execution - */ - executed_history: string[]; - /** - * Results - * @description The results of node executions - */ - results: { - [key: string]: components["schemas"]["BooleanCollectionOutput"] | components["schemas"]["BooleanOutput"] | components["schemas"]["BoundingBoxCollectionOutput"] | components["schemas"]["BoundingBoxOutput"] | components["schemas"]["CLIPOutput"] | components["schemas"]["CLIPSkipInvocationOutput"] | components["schemas"]["CalculateImageTilesOutput"] | components["schemas"]["CogView4ConditioningOutput"] | components["schemas"]["CogView4ModelLoaderOutput"] | components["schemas"]["CollectInvocationOutput"] | components["schemas"]["ColorCollectionOutput"] | components["schemas"]["ColorOutput"] | components["schemas"]["ConditioningCollectionOutput"] | components["schemas"]["ConditioningOutput"] | components["schemas"]["ControlOutput"] | components["schemas"]["DenoiseMaskOutput"] | components["schemas"]["FaceMaskOutput"] | components["schemas"]["FaceOffOutput"] | components["schemas"]["FloatCollectionOutput"] | components["schemas"]["FloatGeneratorOutput"] | components["schemas"]["FloatOutput"] | components["schemas"]["Flux2KleinLoRALoaderOutput"] | components["schemas"]["Flux2KleinModelLoaderOutput"] | components["schemas"]["FluxConditioningCollectionOutput"] | components["schemas"]["FluxConditioningOutput"] | components["schemas"]["FluxControlLoRALoaderOutput"] | components["schemas"]["FluxControlNetOutput"] | components["schemas"]["FluxFillOutput"] | components["schemas"]["FluxKontextOutput"] | components["schemas"]["FluxLoRALoaderOutput"] | components["schemas"]["FluxModelLoaderOutput"] | components["schemas"]["FluxReduxOutput"] | components["schemas"]["GradientMaskOutput"] | components["schemas"]["IPAdapterOutput"] | components["schemas"]["IdealSizeOutput"] | components["schemas"]["ImageCollectionOutput"] | components["schemas"]["ImageGeneratorOutput"] | components["schemas"]["ImageOutput"] | components["schemas"]["ImagePanelCoordinateOutput"] | components["schemas"]["IntegerCollectionOutput"] | components["schemas"]["IntegerGeneratorOutput"] | components["schemas"]["IntegerOutput"] | components["schemas"]["IterateInvocationOutput"] | components["schemas"]["LatentsCollectionOutput"] | components["schemas"]["LatentsMetaOutput"] | components["schemas"]["LatentsOutput"] | components["schemas"]["LoRALoaderOutput"] | components["schemas"]["LoRASelectorOutput"] | components["schemas"]["MDControlListOutput"] | components["schemas"]["MDIPAdapterListOutput"] | components["schemas"]["MDT2IAdapterListOutput"] | components["schemas"]["MaskOutput"] | components["schemas"]["MetadataItemOutput"] | components["schemas"]["MetadataOutput"] | components["schemas"]["MetadataToLorasCollectionOutput"] | components["schemas"]["MetadataToModelOutput"] | components["schemas"]["MetadataToSDXLModelOutput"] | components["schemas"]["ModelIdentifierOutput"] | components["schemas"]["ModelLoaderOutput"] | components["schemas"]["NoiseOutput"] | components["schemas"]["PBRMapsOutput"] | components["schemas"]["PairTileImageOutput"] | components["schemas"]["PromptTemplateOutput"] | components["schemas"]["SD3ConditioningOutput"] | components["schemas"]["SDXLLoRALoaderOutput"] | components["schemas"]["SDXLModelLoaderOutput"] | components["schemas"]["SDXLRefinerModelLoaderOutput"] | components["schemas"]["SchedulerOutput"] | components["schemas"]["Sd3ModelLoaderOutput"] | components["schemas"]["SeamlessModeOutput"] | components["schemas"]["String2Output"] | components["schemas"]["StringCollectionOutput"] | components["schemas"]["StringGeneratorOutput"] | components["schemas"]["StringOutput"] | components["schemas"]["StringPosNegOutput"] | components["schemas"]["T2IAdapterOutput"] | components["schemas"]["TileToPropertiesOutput"] | components["schemas"]["UNetOutput"] | components["schemas"]["VAEOutput"] | components["schemas"]["ZImageConditioningOutput"] | components["schemas"]["ZImageControlOutput"] | components["schemas"]["ZImageLoRALoaderOutput"] | components["schemas"]["ZImageModelLoaderOutput"]; - }; - /** - * Errors - * @description Errors raised when executing nodes - */ - errors: { - [key: string]: string; - }; + FaceMaskOutput: { + /** @description The output image */ + image: components["schemas"]["ImageField"]; /** - * Prepared Source Mapping - * @description The map of prepared nodes to original graph nodes + * Width + * @description The width of the image in pixels */ - prepared_source_mapping: { - [key: string]: string; - }; + width: number; /** - * Source Prepared Mapping - * @description The map of original graph nodes to prepared nodes + * Height + * @description The height of the image in pixels */ - source_prepared_mapping: { - [key: string]: string[]; - }; - /** Ready Order */ - ready_order?: string[]; + height: number; /** - * Indegree - * @description Remaining unmet input count for exec nodes + * type + * @default face_mask_output + * @constant */ - indegree?: { - [key: string]: number; - }; + type: "face_mask_output"; + /** @description The output mask */ + mask: components["schemas"]["ImageField"]; }; /** - * Grounding DINO (Text Prompt Object Detection) - * @description Runs a Grounding DINO model. Performs zero-shot bounding-box object detection from a text prompt. + * FaceOff + * @description Bound, extract, and mask a face from an image using MediaPipe detection */ - GroundingDinoInvocation: { + FaceOffInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10841,40 +11579,112 @@ export type components = { */ use_cache?: boolean; /** - * Model - * @description The Grounding DINO model to use. + * @description Image for face detection * @default null */ - model?: ("grounding-dino-tiny" | "grounding-dino-base") | null; + image?: components["schemas"]["ImageField"] | null; /** - * Prompt - * @description The prompt describing the object to segment. - * @default null + * Face Id + * @description The face ID to process, numbered from 0. Multiple faces not supported. Find a face's ID with FaceIdentifier node. + * @default 0 */ - prompt?: string | null; + face_id?: number; /** - * @description The image to segment. - * @default null + * Minimum Confidence + * @description Minimum confidence for face detection (lower if detection is failing) + * @default 0.5 */ - image?: components["schemas"]["ImageField"] | null; + minimum_confidence?: number; /** - * Detection Threshold - * @description The detection threshold for the Grounding DINO model. All detected bounding boxes with scores above this threshold will be returned. - * @default 0.3 + * X Offset + * @description X-axis offset of the mask + * @default 0 */ - detection_threshold?: number; + x_offset?: number; + /** + * Y Offset + * @description Y-axis offset of the mask + * @default 0 + */ + y_offset?: number; + /** + * Padding + * @description All-axis padding around the mask in pixels + * @default 0 + */ + padding?: number; + /** + * Chunk + * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. + * @default false + */ + chunk?: boolean; /** * type - * @default grounding_dino + * @default face_off * @constant */ - type: "grounding_dino"; + type: "face_off"; }; /** - * HED Edge Detection - * @description Geneartes an edge map using the HED (softedge) model. + * FaceOffOutput + * @description Base class for FaceOff Output */ - HEDEdgeDetectionInvocation: { + FaceOffOutput: { + /** @description The output image */ + image: components["schemas"]["ImageField"]; + /** + * Width + * @description The width of the image in pixels + */ + width: number; + /** + * Height + * @description The height of the image in pixels + */ + height: number; + /** + * type + * @default face_off_output + * @constant + */ + type: "face_off_output"; + /** @description The output mask */ + mask: components["schemas"]["ImageField"]; + /** + * X + * @description The x coordinate of the bounding box's left side + */ + x: number; + /** + * Y + * @description The y coordinate of the bounding box's top side + */ + y: number; + }; + /** + * FieldKind + * @description The kind of field. + * - `Input`: An input field on a node. + * - `Output`: An output field on a node. + * - `Internal`: A field which is treated as an input, but cannot be used in node definitions. Metadata is + * one example. It is provided to nodes via the WithMetadata class, and we want to reserve the field name + * "metadata" for this on all nodes. `FieldKind` is used to short-circuit the field name validation logic, + * allowing "metadata" for that field. + * - `NodeAttribute`: The field is a node attribute. These are fields which are not inputs or outputs, + * but which are used to store information about the node. For example, the `id` and `type` fields are node + * attributes. + * + * The presence of this in `json_schema_extra["field_kind"]` is used when initializing node schemas on app + * startup, and when generating the OpenAPI schema for the workflow editor. + * @enum {string} + */ + FieldKind: "input" | "output" | "internal" | "node_attribute"; + /** + * FilmGrain + * @description Adds film grain to an image + */ + FilmGrainInvocation: { /** * @description The board to save the image to * @default null @@ -10903,62 +11713,68 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The image to add film grain to * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Scribble - * @description Whether or not to use scribble mode - * @default false + * Amount 1 + * @description Amount of the first noise layer + * @default 100 */ - scribble?: boolean; + amount_1?: number; /** - * type - * @default hed_edge_detection - * @constant + * Amount 2 + * @description Amount of the second noise layer + * @default 50 */ - type: "hed_edge_detection"; - }; - /** - * HFModelSource - * @description A HuggingFace repo_id with optional variant, sub-folder(s) and access token. - * Note that the variant option, if not provided to the constructor, will default to fp16, which is - * what people (almost) always want. - * - * The subfolder can be a single path or multiple paths joined by '+' (e.g., "text_encoder+tokenizer"). - * When multiple subfolders are specified, all of them will be downloaded and combined into the model directory. - */ - HFModelSource: { - /** Repo Id */ - repo_id: string; - /** @default fp16 */ - variant?: components["schemas"]["ModelRepoVariant"] | null; - /** Subfolder */ - subfolder?: string | null; - /** Access Token */ - access_token?: string | null; + amount_2?: number; /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} + * Seed 1 + * @description The first seed to use (omit for random) + * @default null */ - type: "hf"; - }; - /** - * HFTokenStatus - * @enum {string} - */ - HFTokenStatus: "valid" | "invalid" | "unknown"; - /** HTTPValidationError */ - HTTPValidationError: { - /** Detail */ - detail?: components["schemas"]["ValidationError"][]; + seed_1?: number | null; + /** + * Seed 2 + * @description The second seed to use (omit for random) + * @default null + */ + seed_2?: number | null; + /** + * Blur 1 + * @description The strength of the first noise blur + * @default 0.5 + */ + blur_1?: number; + /** + * Blur 2 + * @description The strength of the second noise blur + * @default 0.5 + */ + blur_2?: number; + /** + * type + * @default film_grain + * @constant + */ + type: "film_grain"; }; /** - * Heuristic Resize - * @description Resize an image using a heuristic method. Preserves edge maps. + * Flatten Histogram (Grayscale) + * @description Scales the values of an L-mode image by scaling them to the full range 0..255 in equal proportions */ - HeuristicResizeInvocation: { + FlattenHistogramMono: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10977,136 +11793,111 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to resize + * @description Single-channel image for which to flatten the histogram * @default null */ image?: components["schemas"]["ImageField"] | null; - /** - * Width - * @description The width to resize to (px) - * @default 512 - */ - width?: number; - /** - * Height - * @description The height to resize to (px) - * @default 512 - */ - height?: number; /** * type - * @default heuristic_resize + * @default flatten_histogram_mono * @constant */ - type: "heuristic_resize"; + type: "flatten_histogram_mono"; }; /** - * HuggingFaceMetadata - * @description Extended metadata fields provided by HuggingFace. + * Float Batch + * @description Create a batched generation, where the workflow is executed once for each float in the batch. */ - HuggingFaceMetadata: { - /** - * Name - * @description model's name - */ - name: string; - /** - * Files - * @description model files and their sizes - */ - files?: components["schemas"]["RemoteModelFile"][]; - /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} - */ - type: "huggingface"; + FloatBatchInvocation: { /** * Id - * @description The HF model id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ id: string; /** - * Api Response - * @description Response from the HF API as stringified JSON + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - api_response?: string | null; + is_intermediate?: boolean; /** - * Is Diffusers - * @description Whether the metadata is for a Diffusers format model - * @default false + * Use Cache + * @description Whether or not to use the cache + * @default true */ - is_diffusers?: boolean; + use_cache?: boolean; /** - * Ckpt Urls - * @description URLs for all checkpoint format models in the metadata + * Batch Group + * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. + * @default None + * @enum {string} */ - ckpt_urls?: string[] | null; - }; - /** HuggingFaceModels */ - HuggingFaceModels: { + batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; /** - * Urls - * @description URLs for all checkpoint format models in the metadata + * Floats + * @description The floats to batch over + * @default null */ - urls: string[] | null; + floats?: number[] | null; /** - * Is Diffusers - * @description Whether the metadata is for a Diffusers format model + * type + * @default float_batch + * @constant */ - is_diffusers: boolean; + type: "float_batch"; }; - /** IPAdapterField */ - IPAdapterField: { + /** + * Float Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option + */ + FloatCollectionIndexInvocation: { /** - * Image - * @description The IP-Adapter image prompt(s). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - image: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][]; - /** @description The IP-Adapter model to use. */ - ip_adapter_model: components["schemas"]["ModelIdentifierField"]; - /** @description The name of the CLIP image encoder model. */ - image_encoder_model: components["schemas"]["ModelIdentifierField"]; + id: string; /** - * Weight - * @description The weight given to the IP-Adapter. - * @default 1 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - weight?: number | number[]; + is_intermediate?: boolean; /** - * Target Blocks - * @description The IP Adapter blocks to apply - * @default [] + * Use Cache + * @description Whether or not to use the cache + * @default false */ - target_blocks?: string[]; + use_cache?: boolean; /** - * Method - * @description Weight apply method - * @default full + * Random + * @description Random Index? + * @default true */ - method?: string; + random?: boolean; /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) * @default 0 */ - begin_step_percent?: number; + index?: number; /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) - * @default 1 + * Collection + * @description float collection + * @default null */ - end_step_percent?: number; + collection?: number[] | null; /** - * @description The bool mask associated with this IP-Adapter. Excluded regions should be set to False, included regions should be set to True. - * @default null + * type + * @default float_collection_index + * @constant */ - mask?: components["schemas"]["TensorField"] | null; + type: "float_collection_index"; }; /** - * IP-Adapter - SD1.5, SDXL - * @description Collects IP-Adapter info to pass to other nodes. + * Float Collection Primitive + * @description A collection of float primitive values */ - IPAdapterInvocation: { + FloatCollectionInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11125,641 +11916,629 @@ export type components = { */ use_cache?: boolean; /** - * Image - * @description The IP-Adapter image prompt(s). - * @default null - */ - image?: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][] | null; - /** - * IP-Adapter Model - * @description The IP-Adapter model. - * @default null - */ - ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Clip Vision Model - * @description CLIP Vision model to use. Overrides model settings. Mandatory for checkpoint models. - * @default ViT-H - * @enum {string} + * Collection + * @description The collection of float values + * @default [] */ - clip_vision_model?: "ViT-H" | "ViT-G" | "ViT-L"; + collection?: number[]; /** - * Weight - * @description The weight given to the IP-Adapter - * @default 1 + * type + * @default float_collection + * @constant */ - weight?: number | number[]; + type: "float_collection"; + }; + /** + * Float Collection Primitive linked + * @description A collection of float primitive values + */ + FloatCollectionLinkedInvocation: { /** - * Method - * @description The method to apply the IP-Adapter - * @default full - * @enum {string} + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - method?: "full" | "style" | "composition" | "style_strong" | "style_precise"; + id: string; /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) - * @default 0 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - begin_step_percent?: number; + is_intermediate?: boolean; /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) - * @default 1 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - end_step_percent?: number; + use_cache?: boolean; /** - * @description A mask defining the region that this IP-Adapter applies to. - * @default null + * Collection + * @description The collection of float values + * @default [] */ - mask?: components["schemas"]["TensorField"] | null; + collection?: number[]; /** * type - * @default ip_adapter + * @default float_collection_linked * @constant */ - type: "ip_adapter"; - }; - /** - * IPAdapterMetadataField - * @description IP Adapter Field, minus the CLIP Vision Encoder model + type: "float_collection_linked"; + /** + * Value + * @description The float value + * @default 0 + */ + value?: number; + }; + /** + * FloatCollectionOutput + * @description Base class for nodes that output a collection of floats */ - IPAdapterMetadataField: { - /** @description The IP-Adapter image prompt. */ - image: components["schemas"]["ImageField"]; - /** @description The IP-Adapter model. */ - ip_adapter_model: components["schemas"]["ModelIdentifierField"]; + FloatCollectionOutput: { /** - * Clip Vision Model - * @description The CLIP Vision model - * @enum {string} + * Collection + * @description The float collection */ - clip_vision_model: "ViT-L" | "ViT-H" | "ViT-G"; + collection: number[]; /** - * Method - * @description Method to apply IP Weights with - * @enum {string} + * type + * @default float_collection_output + * @constant */ - method: "full" | "style" | "composition" | "style_strong" | "style_precise"; + type: "float_collection_output"; + }; + /** + * Float Collection Toggle + * @description Allows boolean selection between two separate float collection inputs + */ + FloatCollectionToggleInvocation: { /** - * Weight - * @description The weight given to the IP-Adapter + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - weight: number | number[]; + id: string; /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - begin_step_percent: number; + is_intermediate?: boolean; /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) + * Use Cache + * @description Whether or not to use the cache + * @default true */ - end_step_percent: number; - }; - /** IPAdapterOutput */ - IPAdapterOutput: { + use_cache?: boolean; /** - * IP-Adapter - * @description IP-Adapter to apply + * Use Second + * @description Use 2nd Input + * @default false */ - ip_adapter: components["schemas"]["IPAdapterField"]; + use_second?: boolean; + /** + * Col1 + * @description First Float Collection Input + * @default null + */ + col1?: number[] | null; + /** + * Col2 + * @description Second Float Collection Input + * @default null + */ + col2?: number[] | null; /** * type - * @default ip_adapter_output + * @default float_collection_toggle * @constant */ - type: "ip_adapter_output"; + type: "float_collection_toggle"; }; /** - * IPAdapterRecallParameter - * @description IP Adapter configuration for recall + * Float Generator + * @description Generated a range of floats for use in a batched generation */ - IPAdapterRecallParameter: { + FloatGenerator: { /** - * Model Name - * @description The name of the IP Adapter model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - model_name: string; + id: string; /** - * Image Name - * @description The filename of the reference image in outputs/images + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - image_name?: string | null; + is_intermediate?: boolean; /** - * Weight - * @description The weight for the IP Adapter - * @default 1 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - weight?: number; + use_cache?: boolean; /** - * Begin Step Percent - * @description When the IP Adapter is first applied (% of total steps) + * Generator Type + * @description The float generator. */ - begin_step_percent?: number | null; + generator: components["schemas"]["FloatGeneratorField"]; /** - * End Step Percent - * @description When the IP Adapter is last applied (% of total steps) + * type + * @default float_generator + * @constant */ - end_step_percent?: number | null; + type: "float_generator"; + }; + /** FloatGeneratorField */ + FloatGeneratorField: Record; + /** + * FloatGeneratorOutput + * @description Base class for nodes that output a collection of floats + */ + FloatGeneratorOutput: { /** - * Method - * @description The IP Adapter method + * Floats + * @description The generated floats */ - method?: ("full" | "style" | "composition") | null; + floats: number[]; /** - * Image Influence - * @description FLUX Redux image influence (if model is flux_redux) + * type + * @default float_generator_output + * @constant */ - image_influence?: ("lowest" | "low" | "medium" | "high" | "highest") | null; + type: "float_generator_output"; }; - /** IPAdapter_Checkpoint_FLUX_Config */ - IPAdapter_Checkpoint_FLUX_Config: { + /** + * Float Primitive + * @description A float primitive value + */ + FloatInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Value + * @description The float value + * @default 0 */ - file_size: number; + value?: number; /** - * Name - * @description Name of the model. + * type + * @default float + * @constant */ - name: string; + type: "float"; + }; + /** + * Float Range + * @description Creates a range + */ + FloatLinearRangeInvocation: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Start + * @description The first value of the range + * @default 5 */ - cover_image: string | null; + start?: number; /** - * Type - * @default ip_adapter - * @constant + * Stop + * @description The last value of the range + * @default 10 */ - type: "ip_adapter"; + stop?: number; /** - * Format - * @default checkpoint - * @constant + * Steps + * @description number of values to interpolate over (including start and stop) + * @default 30 */ - format: "checkpoint"; + steps?: number; /** - * Base - * @default flux + * type + * @default float_range * @constant */ - base: "flux"; + type: "float_range"; }; - /** IPAdapter_Checkpoint_SD1_Config */ - IPAdapter_Checkpoint_SD1_Config: { + /** + * Float Math + * @description Performs floating point math. + */ + FloatMathInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. + * Operation + * @description The operation to perform + * @default ADD + * @enum {string} */ - name: string; + operation?: "ADD" | "SUB" | "MUL" | "DIV" | "EXP" | "ABS" | "SQRT" | "MIN" | "MAX"; /** - * Description - * @description Model description + * A + * @description The first number + * @default 1 */ - description: string | null; + a?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * B + * @description The second number + * @default 1 */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + b?: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * type + * @default float_math + * @constant */ - source_api_response: string | null; + type: "float_math"; + }; + /** + * FloatOutput + * @description Base class for nodes that output a single float + */ + FloatOutput: { /** - * Cover Image - * @description Url for image to preview model + * Value + * @description The output float */ - cover_image: string | null; + value: number; /** - * Type - * @default ip_adapter + * type + * @default float_output * @constant */ - type: "ip_adapter"; + type: "float_output"; + }; + /** + * Float To Integer + * @description Rounds a float number to (a multiple of) an integer. + */ + FloatToIntegerInvocation: { /** - * Format - * @default checkpoint - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - format: "checkpoint"; + id: string; /** - * Base - * @default sd-1 - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - base: "sd-1"; - }; - /** IPAdapter_Checkpoint_SD2_Config */ - IPAdapter_Checkpoint_SD2_Config: { + is_intermediate?: boolean; /** - * Key - * @description A unique key for this model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - key: string; + use_cache?: boolean; /** - * Hash - * @description The hash of the model file(s). + * Value + * @description The value to round + * @default 0 */ - hash: string; + value?: number; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Multiple of + * @description The multiple to round to + * @default 1 */ - path: string; + multiple?: number; /** - * File Size - * @description The size of the model in bytes. + * Method + * @description The method to use for rounding + * @default Nearest + * @enum {string} */ - file_size: number; + method?: "Nearest" | "Floor" | "Ceiling" | "Truncate"; /** - * Name - * @description Name of the model. + * type + * @default float_to_int + * @constant */ - name: string; + type: "float_to_int"; + }; + /** + * Float Toggle + * @description Allows boolean selection between two separate float inputs + */ + FloatToggleInvocation: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Second + * @description Use 2nd Input + * @default false */ - cover_image: string | null; + use_second?: boolean; /** - * Type - * @default ip_adapter - * @constant + * Float1 + * @description First Float Input + * @default 0 */ - type: "ip_adapter"; + float1?: number; /** - * Format - * @default checkpoint - * @constant + * Float2 + * @description Second Float Input + * @default 0 */ - format: "checkpoint"; + float2?: number; /** - * Base - * @default sd-2 + * type + * @default float_toggle * @constant */ - base: "sd-2"; + type: "float_toggle"; }; - /** IPAdapter_Checkpoint_SDXL_Config */ - IPAdapter_Checkpoint_SDXL_Config: { + /** + * Floats To Strings + * @description Converts a float or collections of floats to a collection of strings + */ + FloatsToStringsInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Floats + * @description float or collection of floats + * @default [] */ - file_size: number; + floats?: number | number[]; /** - * Name - * @description Name of the model. + * type + * @default floats_to_strings + * @constant */ - name: string; + type: "floats_to_strings"; + }; + /** + * FLUX2 Denoise + * @description Run denoising process with a FLUX.2 Klein transformer model. + * + * This node is designed for FLUX.2 Klein models which use Qwen3 as the text encoder. + * It does not support ControlNet, IP-Adapters, or regional prompting. + */ + Flux2DenoiseInvocation: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * @description Latents tensor + * @default null */ - cover_image: string | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * Type - * @default ip_adapter - * @constant + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - type: "ip_adapter"; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Format - * @default checkpoint - * @constant + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - format: "checkpoint"; + denoising_start?: number; /** - * Base - * @default sdxl - * @constant + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - base: "sdxl"; - }; - /** IPAdapter_InvokeAI_SD1_Config */ - IPAdapter_InvokeAI_SD1_Config: { + denoising_end?: number; /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - /** - * Format - * @default invokeai - * @constant - */ - format: "invokeai"; - /** Image Encoder Model Id */ - image_encoder_model_id: string; - /** - * Base - * @default sd-1 - * @constant + * Add Noise + * @description Add noise based on denoising start. + * @default true */ - base: "sd-1"; - }; - /** IPAdapter_InvokeAI_SD2_Config */ - IPAdapter_InvokeAI_SD2_Config: { + add_noise?: boolean; /** - * Key - * @description A unique key for this model. + * Transformer + * @description Flux model (Transformer) to load + * @default null */ - key: string; + transformer?: components["schemas"]["TransformerField"] | null; /** - * Hash - * @description The hash of the model file(s). + * @description Positive conditioning tensor + * @default null */ - hash: string; + positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. + * @default null */ - path: string; + negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | null; /** - * File Size - * @description The size of the model in bytes. + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 1 */ - file_size: number; + cfg_scale?: number; /** - * Name - * @description Name of the model. + * Width + * @description Width of the generated image. + * @default 1024 */ - name: string; + width?: number; /** - * Description - * @description Model description + * Height + * @description Height of the generated image. + * @default 1024 */ - description: string | null; + height?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Num Steps + * @description Number of diffusion steps. Use 4 for distilled models, 28+ for base models. + * @default 4 */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + num_steps?: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Scheduler + * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. + * @default euler + * @enum {string} */ - source_api_response: string | null; + scheduler?: "euler" | "heun" | "lcm"; /** - * Cover Image - * @description Url for image to preview model + * Seed + * @description Randomness seed for reproducibility. + * @default 0 */ - cover_image: string | null; + seed?: number; /** - * Type - * @default ip_adapter - * @constant + * @description FLUX.2 VAE model (required for BN statistics). + * @default null */ - type: "ip_adapter"; + vae?: components["schemas"]["VAEField"] | null; /** - * Format - * @default invokeai - * @constant + * Reference Images + * @description FLUX Kontext conditioning (reference images for multi-reference image editing). + * @default null */ - format: "invokeai"; - /** Image Encoder Model Id */ - image_encoder_model_id: string; + kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; /** - * Base - * @default sd-2 + * type + * @default flux2_denoise * @constant */ - base: "sd-2"; + type: "flux2_denoise"; }; - /** IPAdapter_InvokeAI_SDXL_Config */ - IPAdapter_InvokeAI_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; + /** + * Apply LoRA Collection - Flux2 Klein + * @description Applies a collection of LoRAs to a FLUX.2 Klein transformer and/or Qwen3 text encoder. + */ + Flux2KleinLoRACollectionLoader: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. + * @default null */ - cover_image: string | null; + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * Type - * @default ip_adapter - * @constant + * Transformer + * @description Transformer + * @default null */ - type: "ip_adapter"; + transformer?: components["schemas"]["TransformerField"] | null; /** - * Format - * @default invokeai - * @constant + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder + * @default null */ - format: "invokeai"; - /** Image Encoder Model Id */ - image_encoder_model_id: string; + qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; /** - * Base - * @default sdxl + * type + * @default flux2_klein_lora_collection_loader * @constant */ - base: "sdxl"; + type: "flux2_klein_lora_collection_loader"; }; /** - * Ideal Size - SD1.5, SDXL - * @description Calculates the ideal size for generation to avoid duplication + * Apply LoRA - Flux2 Klein + * @description Apply a LoRA model to a FLUX.2 Klein transformer and/or Qwen3 text encoder. */ - IdealSizeInvocation: { + Flux2KleinLoRALoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11778,62 +12557,74 @@ export type components = { */ use_cache?: boolean; /** - * Width - * @description Final image width - * @default 1024 + * LoRA + * @description LoRA model to load + * @default null */ - width?: number; + lora?: components["schemas"]["ModelIdentifierField"] | null; /** - * Height - * @description Final image height - * @default 576 + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 */ - height?: number; + weight?: number; /** - * @description UNet (scheduler, LoRAs) + * Transformer + * @description Transformer * @default null */ - unet?: components["schemas"]["UNetField"] | null; + transformer?: components["schemas"]["TransformerField"] | null; /** - * Multiplier - * @description Amount to multiply the model's dimensions by when calculating the ideal size (may result in initial generation artifacts if too large) - * @default 1 + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder + * @default null */ - multiplier?: number; + qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; /** * type - * @default ideal_size + * @default flux2_klein_lora_loader * @constant */ - type: "ideal_size"; + type: "flux2_klein_lora_loader"; }; /** - * IdealSizeOutput - * @description Base class for invocations that output an image + * Flux2KleinLoRALoaderOutput + * @description FLUX.2 Klein LoRA Loader Output */ - IdealSizeOutput: { + Flux2KleinLoRALoaderOutput: { /** - * Width - * @description The ideal width of the image (in pixels) + * Transformer + * @description Transformer + * @default null */ - width: number; + transformer: components["schemas"]["TransformerField"] | null; /** - * Height - * @description The ideal height of the image (in pixels) + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder + * @default null */ - height: number; + qwen3_encoder: components["schemas"]["Qwen3EncoderField"] | null; /** * type - * @default ideal_size_output + * @default flux2_klein_lora_loader_output * @constant */ - type: "ideal_size_output"; + type: "flux2_klein_lora_loader_output"; }; /** - * Image Batch - * @description Create a batched generation, where the workflow is executed once for each image in the batch. + * Main Model - Flux2 Klein + * @description Loads a Flux2 Klein model, outputting its submodels. + * + * Flux2 Klein uses Qwen3 as the text encoder instead of CLIP+T5. + * It uses a 32-channel VAE (AutoencoderKLFlux2) instead of the 16-channel FLUX.1 VAE. + * + * When using a Diffusers format model, both VAE and Qwen3 encoder are extracted + * automatically from the main model. You can override with standalone models: + * - Transformer: Always from Flux2 Klein main model + * - VAE: From main model (Diffusers) or standalone VAE + * - Qwen3 Encoder: From main model (Diffusers) or standalone Qwen3 model */ - ImageBatchInvocation: { + Flux2KleinModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11852,40 +12643,84 @@ export type components = { */ use_cache?: boolean; /** - * Batch Group - * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. - * @default None - * @enum {string} + * Transformer + * @description Flux model (Transformer) to load */ - batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; + model: components["schemas"]["ModelIdentifierField"]; /** - * Images - * @description The images to batch over + * VAE + * @description Standalone VAE model. Flux2 Klein uses the same VAE as FLUX (16-channel). If not provided, VAE will be loaded from the Qwen3 Source model. * @default null */ - images?: components["schemas"]["ImageField"][] | null; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Qwen3 Encoder + * @description Standalone Qwen3 Encoder model. If not provided, encoder will be loaded from the Qwen3 Source model. + * @default null + */ + qwen3_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Qwen3 Source (Diffusers) + * @description Diffusers Flux2 Klein model to extract VAE and/or Qwen3 encoder from. Use this if you don't have separate VAE/Qwen3 models. Ignored if both VAE and Qwen3 Encoder are provided separately. + * @default null + */ + qwen3_source_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Max Seq Length + * @description Max sequence length for the Qwen3 encoder. + * @default 512 + * @enum {integer} + */ + max_seq_len?: 256 | 512; /** * type - * @default image_batch + * @default flux2_klein_model_loader * @constant */ - type: "image_batch"; + type: "flux2_klein_model_loader"; }; /** - * Blur Image - * @description Blurs an image + * Flux2KleinModelLoaderOutput + * @description Flux2 Klein model loader output. */ - ImageBlurInvocation: { + Flux2KleinModelLoaderOutput: { /** - * @description The board to save the image to - * @default null + * Transformer + * @description Transformer */ - board?: components["schemas"]["BoardField"] | null; + transformer: components["schemas"]["TransformerField"]; /** - * @description Optional metadata to be saved with the image - * @default null + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder */ - metadata?: components["schemas"]["MetadataField"] | null; + qwen3_encoder: components["schemas"]["Qwen3EncoderField"]; + /** + * VAE + * @description VAE + */ + vae: components["schemas"]["VAEField"]; + /** + * Max Seq Length + * @description The max sequence length for the Qwen3 encoder. + * @enum {integer} + */ + max_seq_len: 256 | 512; + /** + * type + * @default flux2_klein_model_loader_output + * @constant + */ + type: "flux2_klein_model_loader_output"; + }; + /** + * Prompt - Flux2 Klein + * @description Encodes and preps a prompt for Flux2 Klein image generation. + * + * Flux2 Klein uses Qwen3 as the text encoder, extracting hidden states from + * layers (9, 18, 27) and stacking them for richer text representations. + * This matches the diffusers Flux2KleinPipeline implementation exactly. + */ + Flux2KleinTextEncoderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11904,47 +12739,41 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to blur + * Prompt + * @description Text prompt to encode. * @default null */ - image?: components["schemas"]["ImageField"] | null; + prompt?: string | null; /** - * Radius - * @description The blur radius - * @default 8 + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder + * @default null */ - radius?: number; + qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; /** - * Blur Type - * @description The type of blur - * @default gaussian - * @enum {string} + * Max Seq Len + * @description Max sequence length for the Qwen3 encoder. + * @default 512 + * @enum {integer} */ - blur_type?: "gaussian" | "box"; + max_seq_len?: 256 | 512; + /** + * @description A mask defining the region that this conditioning prompt applies to. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default img_blur + * @default flux2_klein_text_encoder * @constant */ - type: "img_blur"; + type: "flux2_klein_text_encoder"; }; /** - * ImageCategory - * @description The category of an image. - * - * - GENERAL: The image is an output, init image, or otherwise an image without a specialized purpose. - * - MASK: The image is a mask image. - * - CONTROL: The image is a ControlNet control image. - * - USER: The image is a user-provide image. - * - OTHER: The image is some other type of image with a specialized purpose. To be used by external nodes. - * @enum {string} - */ - ImageCategory: "general" | "mask" | "control" | "user" | "other"; - /** - * Extract Image Channel - * @description Gets a channel from an image. + * Latents to Image - FLUX2 + * @description Generates an image from latents using FLUX.2 Klein's 32-channel VAE. */ - ImageChannelInvocation: { + Flux2VaeDecodeInvocation: { /** * @description The board to save the image to * @default null @@ -11973,39 +12802,27 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to get the channel from + * @description Latents tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * Channel - * @description The channel to get - * @default A - * @enum {string} + * @description VAE + * @default null */ - channel?: "A" | "R" | "G" | "B"; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default img_chan + * @default flux2_vae_decode * @constant */ - type: "img_chan"; + type: "flux2_vae_decode"; }; /** - * Multiply Image Channel - * @description Scale a specific color channel of an image. + * Image to Latents - FLUX2 + * @description Encodes an image into latents using FLUX.2 Klein's 32-channel VAE. */ - ImageChannelMultiplyInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + Flux2VaeEncodeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12024,50 +12841,34 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to adjust + * @description The image to encode. * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Channel - * @description Which channel to adjust + * @description VAE * @default null */ - channel?: ("Red (RGBA)" | "Green (RGBA)" | "Blue (RGBA)" | "Alpha (RGBA)" | "Cyan (CMYK)" | "Magenta (CMYK)" | "Yellow (CMYK)" | "Black (CMYK)" | "Hue (HSV)" | "Saturation (HSV)" | "Value (HSV)" | "Luminosity (LAB)" | "A (LAB)" | "B (LAB)" | "Y (YCbCr)" | "Cb (YCbCr)" | "Cr (YCbCr)") | null; - /** - * Scale - * @description The amount to scale the channel by. - * @default 1 - */ - scale?: number; - /** - * Invert Channel - * @description Invert the channel after scaling - * @default false - */ - invert_channel?: boolean; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default img_channel_multiply + * @default flux2_vae_encode * @constant */ - type: "img_channel_multiply"; + type: "flux2_vae_encode"; }; /** - * Offset Image Channel - * @description Add or subtract a value from a specific color channel of an image. + * Flux2VariantType + * @description FLUX.2 model variants. + * @enum {string} */ - ImageChannelOffsetInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + Flux2VariantType: "klein_4b" | "klein_9b" | "klein_9b_base"; + /** + * Flux Conditioning Blend + * @description Performs a blend between two FLUX Conditioning objects using either direct SLERP + * or an advanced method that separates magnitude and direction. + */ + FluxConditioningBlendInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12086,34 +12887,53 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to adjust + * @description The first FLUX Conditioning object. * @default null */ - image?: components["schemas"]["ImageField"] | null; + conditioning_1?: components["schemas"]["FluxConditioningField"] | null; /** - * Channel - * @description Which channel to adjust + * @description The second FLUX Conditioning object. * @default null */ - channel?: ("Red (RGBA)" | "Green (RGBA)" | "Blue (RGBA)" | "Alpha (RGBA)" | "Cyan (CMYK)" | "Magenta (CMYK)" | "Yellow (CMYK)" | "Black (CMYK)" | "Hue (HSV)" | "Saturation (HSV)" | "Value (HSV)" | "Luminosity (LAB)" | "A (LAB)" | "B (LAB)" | "Y (YCbCr)" | "Cb (YCbCr)" | "Cr (YCbCr)") | null; + conditioning_2?: components["schemas"]["FluxConditioningField"] | null; /** - * Offset - * @description The amount to adjust the channel by - * @default 0 + * Alpha + * @description Interpolation factor (0.0 for conditioning_1, 1.0 for conditioning_2). + * @default 0.5 */ - offset?: number; + alpha?: number; + /** + * Use Magnitude Separation + * @description If True, uses magnitude separation (SLERP for direction, LERP for magnitude); otherwise, uses direct SLERP. + * @default false + */ + use_magnitude_separation?: boolean; /** * type - * @default img_channel_offset + * @default flux_conditioning_blend * @constant */ - type: "img_channel_offset"; + type: "flux_conditioning_blend"; }; /** - * Image Collection Primitive - * @description A collection of image primitive values + * FluxConditioningBlendOutput + * @description Output for the blended Flux Conditionings. */ - ImageCollectionInvocation: { + FluxConditioningBlendOutput: { + /** @description The interpolated Flux Conditioning */ + conditioning: components["schemas"]["FluxConditioningField"]; + /** + * type + * @default flux_conditioning_blend_output + * @constant + */ + type: "flux_conditioning_blend_output"; + }; + /** + * Flux Conditioning Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option + */ + FluxConditioningCollectionIndexInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12128,54 +12948,39 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * Collection - * @description The collection of image values - * @default null + * Random + * @description Random Index? + * @default true */ - collection?: components["schemas"]["ImageField"][] | null; + random?: boolean; /** - * type - * @default image_collection - * @constant + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 */ - type: "image_collection"; - }; - /** - * ImageCollectionOutput - * @description Base class for nodes that output a collection of images - */ - ImageCollectionOutput: { + index?: number; /** - * Collection - * @description The output images + * FLUX Conditionings + * @description Conditioning tensor + * @default [] */ - collection: components["schemas"]["ImageField"][]; + collection?: components["schemas"]["FluxConditioningField"][]; /** * type - * @default image_collection_output + * @default flux_conditioning_index * @constant */ - type: "image_collection_output"; + type: "flux_conditioning_index"; }; /** - * Convert Image Mode - * @description Converts an image to a different mode. + * Flux Conditioning Collection Primitive + * @description A collection of flux conditioning tensor primitive values */ - ImageConvertInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FluxConditioningCollectionInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12194,39 +12999,23 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to convert - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Mode - * @description The mode to convert to - * @default L - * @enum {string} + * FLUX Conditionings + * @description Conditioning tensor + * @default [] */ - mode?: "L" | "RGB" | "RGBA" | "CMYK" | "YCbCr" | "LAB" | "HSV" | "I" | "F"; + conditioning?: components["schemas"]["FluxConditioningField"][]; /** * type - * @default img_conv + * @default flux_conditioning_collection * @constant */ - type: "img_conv"; + type: "flux_conditioning_collection"; }; /** - * Crop Image - * @description Crops an image to a specified box. The box can be outside of the image. + * Flux Conditioning Collection join + * @description Join a flux conditioning tensor or collections into a single collection of flux conditioning tensors */ - ImageCropInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FluxConditioningCollectionJoinInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12245,137 +13034,145 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to crop + * FLUX Text Encoder Conditioning or Collection + * @description Conditioning tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * X - * @description The left x coordinate of the crop rectangle - * @default 0 - */ - x?: number; - /** - * Y - * @description The top y coordinate of the crop rectangle - * @default 0 - */ - y?: number; - /** - * Width - * @description The width of the crop rectangle - * @default 512 - */ - width?: number; + conditionings_a?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * Height - * @description The height of the crop rectangle - * @default 512 + * FLUX Text Encoder Conditioning or Collection + * @description Conditioning tensor + * @default null */ - height?: number; + conditionings_b?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** * type - * @default img_crop + * @default flux_conditioning_collection_join * @constant */ - type: "img_crop"; + type: "flux_conditioning_collection_join"; }; /** - * ImageDTO - * @description Deserialized image record, enriched for the frontend. + * FluxConditioningCollectionOutput + * @description Base class for nodes that output a collection of conditioning tensors */ - ImageDTO: { - /** - * Image Name - * @description The unique name of the image. - */ - image_name: string; + FluxConditioningCollectionOutput: { /** - * Image Url - * @description The URL of the image. + * Collection + * @description The output conditioning tensors */ - image_url: string; + collection: components["schemas"]["FluxConditioningField"][]; /** - * Thumbnail Url - * @description The URL of the image's thumbnail. + * type + * @default flux_conditioning_collection_output + * @constant */ - thumbnail_url: string; - /** @description The type of the image. */ - image_origin: components["schemas"]["ResourceOrigin"]; - /** @description The category of the image. */ - image_category: components["schemas"]["ImageCategory"]; + type: "flux_conditioning_collection_output"; + }; + /** + * FluxConditioningDeltaAndAugmentedOutput + * @description Output for the Conditioning Delta and Augmented Conditioning. + */ + FluxConditioningDeltaAndAugmentedOutput: { + /** @description The augmented conditioning (base + delta, or just delta) */ + augmented_conditioning: components["schemas"]["FluxConditioningField"]; + /** @description The resulting conditioning delta (feature - reference) */ + delta_conditioning: components["schemas"]["FluxConditioningField"]; /** - * Width - * @description The width of the image in px. + * type + * @default conditioning_delta_and_augmented_output + * @constant */ - width: number; + type: "conditioning_delta_and_augmented_output"; + }; + /** + * Flux Conditioning Delta + * @description Calculates the delta between feature and reference conditionings, + * and optionally augments a base conditioning with this delta. + * If reference conditioning is omitted, it will be treated as zero tensors. + */ + FluxConditioningDeltaAugmentationInvocation: { /** - * Height - * @description The height of the image in px. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - height: number; + id: string; /** - * Created At - * @description The created timestamp of the image. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - created_at: string; + is_intermediate?: boolean; /** - * Updated At - * @description The updated timestamp of the image. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - updated_at: string; + use_cache?: boolean; /** - * Deleted At - * @description The deleted timestamp of the image. + * Feature Conditioning + * @description Feature Conditioning (single or list) for delta calculation. If a list, it will be averaged. + * @default null */ - deleted_at?: string | null; + feature_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * Is Intermediate - * @description Whether this is an intermediate image. + * Reference Conditioning + * @description Reference Conditioning (single or list) for delta calculation. If a list, it will be averaged. If omitted, zero tensors will be used as reference. + * @default null */ - is_intermediate: boolean; + reference_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * Session Id - * @description The session ID that generated this image, if it is a generated image. + * @description Optional Base Conditioning to which the delta will be added. If not provided, Augmented Conditioning will be the Delta. + * @default null */ - session_id?: string | null; + base_conditioning?: components["schemas"]["FluxConditioningField"] | null; /** - * Node Id - * @description The node ID that generated this image, if it is a generated image. + * Base Scale + * @description Scalar to multiply the base conditioning when augmenting. + * @default 1 */ - node_id?: string | null; + base_scale?: number; /** - * Starred - * @description Whether this image is starred. + * Delta Scale + * @description Scalar to multiply the delta when augmenting the base conditioning. + * @default 1 */ - starred: boolean; + delta_scale?: number; /** - * Has Workflow - * @description Whether this image has a workflow. + * Scale Delta Output + * @description If true, the delta output will also be scaled by the delta_scale. + * @default false */ - has_workflow: boolean; + scale_delta_output?: boolean; /** - * Board Id - * @description The id of the board the image belongs to, if one exists. + * type + * @default flux_conditioning_delta_augmentation + * @constant */ - board_id?: string | null; + type: "flux_conditioning_delta_augmentation"; }; /** - * ImageField - * @description An image primitive field + * FluxConditioningField + * @description A conditioning tensor primitive value */ - ImageField: { + FluxConditioningField: { /** - * Image Name - * @description The name of the image + * Conditioning Name + * @description The name of conditioning tensor */ - image_name: string; + conditioning_name: string; + /** + * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; }; /** - * Image Generator - * @description Generated a collection of images for use in a batched generation + * Flux Conditioning List + * @description Takes multiple optional Flux Conditioning inputs and outputs them as a single + * ordered list. Missing (None) inputs are gracefully handled. */ - ImageGenerator: { + FluxConditioningListInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12394,51 +13191,65 @@ export type components = { */ use_cache?: boolean; /** - * Generator Type - * @description The image generator. + * @description First optional Flux Conditioning input. + * @default null */ - generator: components["schemas"]["ImageGeneratorField"]; + conditioning_1?: components["schemas"]["FluxConditioningField"] | null; /** - * type - * @default image_generator - * @constant + * @description Second optional Flux Conditioning input. + * @default null */ - type: "image_generator"; - }; - /** ImageGeneratorField */ - ImageGeneratorField: Record; - /** - * ImageGeneratorOutput - * @description Base class for nodes that output a collection of boards - */ - ImageGeneratorOutput: { + conditioning_2?: components["schemas"]["FluxConditioningField"] | null; /** - * Images - * @description The generated images + * @description Third optional Flux Conditioning input. + * @default null */ - images: components["schemas"]["ImageField"][]; + conditioning_3?: components["schemas"]["FluxConditioningField"] | null; + /** + * @description Fourth optional Flux Conditioning input. + * @default null + */ + conditioning_4?: components["schemas"]["FluxConditioningField"] | null; + /** + * @description Fifth optional Flux Conditioning input. + * @default null + */ + conditioning_5?: components["schemas"]["FluxConditioningField"] | null; + /** + * @description Sixth optional Flux Conditioning input. + * @default null + */ + conditioning_6?: components["schemas"]["FluxConditioningField"] | null; /** * type - * @default image_generator_output + * @default flux_conditioning_list * @constant */ - type: "image_generator_output"; + type: "flux_conditioning_list"; }; /** - * Adjust Image Hue - * @description Adjusts the Hue of an image. + * FluxConditioningListOutput + * @description Output for the Flux Conditioning List node, providing an ordered list + * of Flux Conditioning objects. */ - ImageHueAdjustmentInvocation: { + FluxConditioningListOutput: { /** - * @description The board to save the image to - * @default null + * Conditioning List + * @description An ordered list of provided Flux Conditioning objects. */ - board?: components["schemas"]["BoardField"] | null; + conditioning_list: components["schemas"]["FluxConditioningField"][]; /** - * @description Optional metadata to be saved with the image - * @default null + * type + * @default flux_conditioning_list_output + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "flux_conditioning_list_output"; + }; + /** + * FLUX Conditioning Math + * @description Performs a Math operation on two FLUX conditionings. + */ + FluxConditioningMathOperationInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12457,38 +13268,77 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to adjust + * @description First FLUX Conditioning (A) * @default null */ - image?: components["schemas"]["ImageField"] | null; + cond_a?: components["schemas"]["FluxConditioningField"] | null; /** - * Hue - * @description The degrees by which to rotate the hue, 0-360 + * @description Second FLUX Conditioning (B) + * @default null + */ + cond_b?: components["schemas"]["FluxConditioningField"] | null; + /** + * Operation + * @description Operation to perform (A op B) + * @default ADD + * @enum {string} + */ + operation?: "ADD" | "SUB" | "MUL" | "DIV" | "APPEND" | "SPV" | "NSPV" | "LERP" | "SLERP"; + /** + * Scale + * @description Scaling factor + * @default 1 + */ + scale?: number; + /** + * Rescale Target Norm + * @description If > 0, rescales the output embeddings to the target max norm. Set to 0 to disable. * @default 0 */ - hue?: number; + rescale_target_norm?: number; /** * type - * @default img_hue_adjust + * @default flux_conditioning_math * @constant */ - type: "img_hue_adjust"; + type: "flux_conditioning_math"; }; /** - * Inverse Lerp Image - * @description Inverse linear interpolation of all pixels of an image + * FluxConditioningOutput + * @description Base class for nodes that output a single conditioning tensor */ - ImageInverseLerpInvocation: { + FluxConditioningOutput: { + /** @description Conditioning tensor */ + conditioning: components["schemas"]["FluxConditioningField"]; /** - * @description The board to save the image to - * @default null + * type + * @default flux_conditioning_output + * @constant */ - board?: components["schemas"]["BoardField"] | null; + type: "flux_conditioning_output"; + }; + /** + * FluxConditioningStoreOutput + * @description Output for the Store Flux Conditioning node. + */ + FluxConditioningStoreOutput: { /** - * @description Optional metadata to be saved with the image - * @default null + * Conditioning Id + * @description Unique identifier for the stored Flux Conditioning */ - metadata?: components["schemas"]["MetadataField"] | null; + conditioning_id: string; + /** + * type + * @default flux_conditioning_store_output + * @constant + */ + type: "flux_conditioning_store_output"; + }; + /** + * Control LoRA - FLUX + * @description LoRA model and Image to use with FLUX transformer generation. + */ + FluxControlLoRALoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12507,34 +13357,52 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to lerp + * Control LoRA + * @description Control LoRA model to load + * @default null + */ + lora?: components["schemas"]["ModelIdentifierField"] | null; + /** + * @description The image to encode. * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Min - * @description The minimum input value - * @default 0 + * Weight + * @description The weight of the LoRA. + * @default 1 */ - min?: number; + weight?: number; /** - * Max - * @description The maximum input value - * @default 255 + * type + * @default flux_control_lora_loader + * @constant */ - max?: number; + type: "flux_control_lora_loader"; + }; + /** + * FluxControlLoRALoaderOutput + * @description Flux Control LoRA Loader Output + */ + FluxControlLoRALoaderOutput: { + /** + * Flux Control LoRA + * @description Control LoRAs to apply on model loading + * @default null + */ + control_lora: components["schemas"]["ControlLoRAField"]; /** * type - * @default img_ilerp + * @default flux_control_lora_loader_output * @constant */ - type: "img_ilerp"; + type: "flux_control_lora_loader_output"; }; /** - * Image Primitive - * @description An image primitive value + * Flux ControlNet Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option */ - ImageInvocation: { + FluxControlNetCollectionIndexInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12549,36 +13417,39 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * @description The image to load - * @default null + * Random + * @description Random Index? + * @default true */ - image?: components["schemas"]["ImageField"] | null; + random?: boolean; + /** + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 + */ + index?: number; + /** + * FLUX ControlNets + * @description FLUX ControlNet Collection + * @default [] + */ + collection?: components["schemas"]["FluxControlNetField"][]; /** * type - * @default image + * @default flux_controlnet_index * @constant */ - type: "image"; + type: "flux_controlnet_index"; }; /** - * Lerp Image - * @description Linear interpolation of all pixels of an image + * FLUX ControlNet Collection Primitive + * @description A collection of flux controlnet primitive values */ - ImageLerpInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FluxControlNetCollectionInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12597,39 +13468,23 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to lerp + * FLUX ControlNet Collection + * @description FLUX ControlNets * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * Min - * @description The minimum output value - * @default 0 - */ - min?: number; - /** - * Max - * @description The maximum output value - * @default 255 - */ - max?: number; + collection?: components["schemas"]["FluxControlNetField"][] | null; /** * type - * @default img_lerp + * @default flux_controlnet_collection * @constant */ - type: "img_lerp"; + type: "flux_controlnet_collection"; }; /** - * Image Mask to Tensor - * @description Convert a mask image to a tensor. Converts the image to grayscale and uses thresholding at the specified value. + * FLUX ControlNet Collection join + * @description Join a flux controlnet tensors or collections into a single collection of flux controlnet tensors */ - ImageMaskToTensorInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FluxControlNetCollectionJoinInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12648,44 +13503,81 @@ export type components = { */ use_cache?: boolean; /** - * @description The mask image to convert. + * FLUX ControlNet or Collection + * @description FLUX ControlNets * @default null */ - image?: components["schemas"]["ImageField"] | null; + controlnets_a?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; /** - * Cutoff - * @description Cutoff (<) - * @default 128 + * FLUX ControlNet or Collection + * @description FLUX ControlNets + * @default null */ - cutoff?: number; + controlnets_b?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; /** - * Invert - * @description Whether to invert the mask. - * @default false + * type + * @default flux_controlnet_collection_join + * @constant */ - invert?: boolean; + type: "flux_controlnet_collection_join"; + }; + /** FluxControlNetCollectionOutput */ + FluxControlNetCollectionOutput: { + /** + * FLUX ControlNet List + * @description ControlNet(s) to apply + */ + collection: components["schemas"]["FluxControlNetField"][]; /** * type - * @default image_mask_to_tensor + * @default flux_controlnet_collection_output * @constant */ - type: "image_mask_to_tensor"; + type: "flux_controlnet_collection_output"; }; - /** - * Multiply Images - * @description Multiplies two images together using `PIL.ImageChops.multiply()`. - */ - ImageMultiplyInvocation: { + /** FluxControlNetField */ + FluxControlNetField: { + /** @description The control image */ + image: components["schemas"]["ImageField"]; + /** @description The ControlNet model to use */ + control_model: components["schemas"]["ModelIdentifierField"]; /** - * @description The board to save the image to - * @default null + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - board?: components["schemas"]["BoardField"] | null; + control_weight?: number | number[]; /** - * @description Optional metadata to be saved with the image - * @default null + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - metadata?: components["schemas"]["MetadataField"] | null; + begin_step_percent?: number; + /** + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; + /** + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} + */ + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + /** + * Instantx Control Mode + * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. + * @default -1 + */ + instantx_control_mode?: number | null; + }; + /** + * FLUX ControlNet + * @description Collect FLUX ControlNet info to pass to other nodes. + */ + FluxControlNetInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12704,102 +13596,58 @@ export type components = { */ use_cache?: boolean; /** - * @description The first image to multiply + * @description The control image * @default null */ - image1?: components["schemas"]["ImageField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description The second image to multiply - * @default null - */ - image2?: components["schemas"]["ImageField"] | null; - /** - * type - * @default img_mul - * @constant - */ - type: "img_mul"; - }; - /** - * Blur NSFW Image - * @description Add blur to NSFW-flagged images - */ - ImageNSFWBlurInvocation: { - /** - * @description The board to save the image to + * @description ControlNet model to load * @default null */ - board?: components["schemas"]["BoardField"] | null; + control_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - metadata?: components["schemas"]["MetadataField"] | null; + control_weight?: number | number[]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - id: string; + begin_step_percent?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - is_intermediate?: boolean; + end_step_percent?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Resize Mode + * @description The resize mode used + * @default just_resize + * @enum {string} */ - use_cache?: boolean; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** - * @description The image to check - * @default null + * Instantx Control Mode + * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. + * @default -1 */ - image?: components["schemas"]["ImageField"] | null; + instantx_control_mode?: number | null; /** * type - * @default img_nsfw + * @default flux_controlnet * @constant */ - type: "img_nsfw"; - }; - /** - * ImageNamesResult - * @description Response containing ordered image names with metadata for optimistic updates. - */ - ImageNamesResult: { - /** - * Image Names - * @description Ordered list of image names - */ - image_names: string[]; - /** - * Starred Count - * @description Number of starred images (when starred_first=True) - */ - starred_count: number; - /** - * Total Count - * @description Total number of images matching the query - */ - total_count: number; + type: "flux_controlnet"; }; /** - * Add Image Noise - * @description Add noise to an image + * FLUX ControlNet-Linked + * @description Collects FLUX ControlNet info to pass to other nodes. */ - ImageNoiseInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FluxControlNetLinkedInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12818,112 +13666,92 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to add noise to + * @description The control image * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description Optional mask determining where to apply noise (black=noise, white=no noise) + * @description ControlNet model to load * @default null */ - mask?: components["schemas"]["ImageField"] | null; + control_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Seed - * @description Seed for random number generation - * @default 0 + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - seed?: number; + control_weight?: number | number[]; /** - * Noise Type - * @description The type of noise to add - * @default gaussian - * @enum {string} + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - noise_type?: "gaussian" | "salt_and_pepper"; + begin_step_percent?: number; /** - * Amount - * @description The amount of noise to add - * @default 0.1 + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - amount?: number; + end_step_percent?: number; /** - * Noise Color - * @description Whether to add colored noise - * @default true + * Resize Mode + * @description The resize mode used + * @default just_resize + * @enum {string} */ - noise_color?: boolean; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** - * Size - * @description The size of the noise points - * @default 1 + * Instantx Control Mode + * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. + * @default -1 */ - size?: number; + instantx_control_mode?: number | null; /** * type - * @default img_noise + * @default flux_controlnet_linked * @constant */ - type: "img_noise"; - }; - /** - * ImageOutput - * @description Base class for nodes that output a single image - */ - ImageOutput: { - /** @description The output image */ - image: components["schemas"]["ImageField"]; + type: "flux_controlnet_linked"; /** - * Width - * @description The width of the image in pixels + * FLUX ControlNet List + * @description FLUX ControlNet List + * @default null */ - width: number; + controlnet_list?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; + }; + /** FluxControlNetListOutput */ + FluxControlNetListOutput: { /** - * Height - * @description The height of the image in pixels + * FLUX ControlNet List + * @description ControlNet(s) to apply */ - height: number; + controlnet_list: components["schemas"]["FluxControlNetField"][]; /** * type - * @default image_output + * @default flux_controlnet_list_output * @constant */ - type: "image_output"; + type: "flux_controlnet_list_output"; }; - /** ImagePanelCoordinateOutput */ - ImagePanelCoordinateOutput: { - /** - * X Left - * @description The left x-coordinate of the panel. - */ - x_left: number; - /** - * Y Top - * @description The top y-coordinate of the panel. - */ - y_top: number; - /** - * Width - * @description The width of the panel. - */ - width: number; - /** - * Height - * @description The height of the panel. - */ - height: number; + /** + * FluxControlNetOutput + * @description FLUX ControlNet info + */ + FluxControlNetOutput: { + /** @description ControlNet(s) to apply */ + control: components["schemas"]["FluxControlNetField"]; /** * type - * @default image_panel_coordinate_output + * @default flux_controlnet_output * @constant */ - type: "image_panel_coordinate_output"; + type: "flux_controlnet_output"; }; /** - * Image Panel Layout - * @description Get the coordinates of a single panel in a grid. (If the full image shape cannot be divided evenly into panels, - * then the grid may not cover the entire image.) + * FLUX Denoise + * @description Run denoising process with a FLUX transformer model. */ - ImagePanelLayoutInvocation: { + FluxDenoiseInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12942,224 +13770,177 @@ export type components = { */ use_cache?: boolean; /** - * Width - * @description The width of the entire grid. + * @description Latents tensor * @default null */ - width?: number | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * Height - * @description The height of the entire grid. + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. * @default null */ - height?: number | null; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Num Cols - * @description The number of columns in the grid. - * @default 1 + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - num_cols?: number; + denoising_start?: number; /** - * Num Rows - * @description The number of rows in the grid. + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps * @default 1 */ - num_rows?: number; + denoising_end?: number; /** - * Panel Col Idx - * @description The column index of the panel to be processed. - * @default 0 + * Add Noise + * @description Add noise based on denoising start. + * @default true */ - panel_col_idx?: number; + add_noise?: boolean; /** - * Panel Row Idx - * @description The row index of the panel to be processed. - * @default 0 + * Transformer + * @description Flux model (Transformer) to load + * @default null */ - panel_row_idx?: number; + transformer?: components["schemas"]["TransformerField"] | null; /** - * type - * @default image_panel_layout - * @constant + * Control LoRA + * @description Control LoRA model to load + * @default null */ - type: "image_panel_layout"; - }; - /** - * Paste Image - * @description Pastes an image into another image. - */ - ImagePasteInvocation: { + control_lora?: components["schemas"]["ControlLoRAField"] | null; /** - * @description The board to save the image to + * Positive Text Conditioning + * @description Positive conditioning tensor * @default null */ - board?: components["schemas"]["BoardField"] | null; + positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * @description Optional metadata to be saved with the image + * Negative Text Conditioning + * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; + negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * @description The base image + * Redux Conditioning + * @description FLUX Redux conditioning tensor. * @default null */ - base_image?: components["schemas"]["ImageField"] | null; + redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; /** - * @description The image to paste + * @description FLUX Fill conditioning. * @default null */ - image?: components["schemas"]["ImageField"] | null; + fill_conditioning?: components["schemas"]["FluxFillConditioningField"] | null; /** - * @description The mask to use when pasting - * @default null + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 1 */ - mask?: components["schemas"]["ImageField"] | null; + cfg_scale?: number | number[]; /** - * X - * @description The left x coordinate at which to paste the image + * CFG Scale Start Step + * @description Index of the first step to apply cfg_scale. Negative indices count backwards from the the last step (e.g. a value of -1 refers to the final step). * @default 0 */ - x?: number; + cfg_scale_start_step?: number; /** - * Y - * @description The top y coordinate at which to paste the image - * @default 0 + * CFG Scale End Step + * @description Index of the last step to apply cfg_scale. Negative indices count backwards from the last step (e.g. a value of -1 refers to the final step). + * @default -1 */ - y?: number; + cfg_scale_end_step?: number; /** - * Crop - * @description Crop to base image dimensions - * @default false + * Width + * @description Width of the generated image. + * @default 1024 */ - crop?: boolean; + width?: number; /** - * type - * @default img_paste - * @constant + * Height + * @description Height of the generated image. + * @default 1024 */ - type: "img_paste"; - }; - /** - * ImageRecordChanges - * @description A set of changes to apply to an image record. - * - * Only limited changes are valid: - * - `image_category`: change the category of an image - * - `session_id`: change the session associated with an image - * - `is_intermediate`: change the image's `is_intermediate` flag - * - `starred`: change whether the image is starred - */ - ImageRecordChanges: { - /** @description The image's new category. */ - image_category?: components["schemas"]["ImageCategory"] | null; + height?: number; /** - * Session Id - * @description The image's new session ID. + * Num Steps + * @description Number of diffusion steps. Recommended values are schnell: 4, dev: 50. + * @default 4 */ - session_id?: string | null; + num_steps?: number; /** - * Is Intermediate - * @description The image's new `is_intermediate` flag. + * Scheduler + * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. + * @default euler + * @enum {string} */ - is_intermediate?: boolean | null; + scheduler?: "euler" | "heun" | "lcm"; /** - * Starred - * @description The image's new `starred` state + * Guidance + * @description The guidance strength. Higher values adhere more strictly to the prompt, and will produce less diverse images. FLUX dev only, ignored for schnell. + * @default 4 */ - starred?: boolean | null; - } & { - [key: string]: unknown; - }; - /** - * Resize Image - * @description Resizes an image to specific dimensions - */ - ImageResizeInvocation: { + guidance?: number; /** - * @description The board to save the image to - * @default null + * Seed + * @description Randomness seed for reproducibility. + * @default 0 */ - board?: components["schemas"]["BoardField"] | null; + seed?: number; /** - * @description Optional metadata to be saved with the image + * Control + * @description ControlNet models. * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + control?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * @description VAE + * @default null */ - is_intermediate?: boolean; + controlnet_vae?: components["schemas"]["VAEField"] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * IP-Adapter + * @description IP-Adapter to apply + * @default null */ - use_cache?: boolean; + ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * @description The image to resize + * Kontext Conditioning + * @description FLUX Kontext conditioning (reference image). * @default null */ - image?: components["schemas"]["ImageField"] | null; + kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; /** - * Width - * @description The width to resize to (px) - * @default 512 + * Dype Preset + * @description DyPE preset for high-resolution generation. 'auto' enables automatically for resolutions > 1536px. 'area' enables automatically based on image area. '4k' uses optimized settings for 4K output. + * @default off + * @enum {string} */ - width?: number; + dype_preset?: "off" | "manual" | "auto" | "area" | "4k"; /** - * Height - * @description The height to resize to (px) - * @default 512 + * Dype Scale + * @description DyPE magnitude (λs). Higher values = stronger extrapolation. Only used when dype_preset is not 'off'. + * @default null */ - height?: number; + dype_scale?: number | null; /** - * Resample Mode - * @description The resampling mode - * @default bicubic - * @enum {string} + * Dype Exponent + * @description DyPE decay speed (λt). Controls transition from low to high frequency detail. Only used when dype_preset is not 'off'. + * @default null */ - resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; + dype_exponent?: number | null; /** * type - * @default img_resize + * @default flux_denoise * @constant */ - type: "img_resize"; + type: "flux_denoise"; }; /** - * Scale Image - * @description Scales an image by a factor + * FLUX Denoise + Metadata + * @description Run denoising process with a FLUX transformer model + metadata. */ - ImageScaleInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + FluxDenoiseLatentsMetaInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -13183,203 +13964,187 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to scale + * @description Latents tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * Scale Factor - * @description The factor by which to scale the image - * @default 2 - */ - scale_factor?: number; - /** - * Resample Mode - * @description The resampling mode - * @default bicubic - * @enum {string} - */ - resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; + latents?: components["schemas"]["LatentsField"] | null; /** - * type - * @default img_scale - * @constant + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - type: "img_scale"; - }; - /** - * Image to Latents - SD1.5, SDXL - * @description Encodes an image into latents. - */ - ImageToLatentsInvocation: { + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - id: string; + denoising_start?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - is_intermediate?: boolean; + denoising_end?: number; /** - * Use Cache - * @description Whether or not to use the cache + * Add Noise + * @description Add noise based on denoising start. * @default true */ - use_cache?: boolean; + add_noise?: boolean; /** - * @description The image to encode + * Transformer + * @description Flux model (Transformer) to load * @default null */ - image?: components["schemas"]["ImageField"] | null; + transformer?: components["schemas"]["TransformerField"] | null; /** - * @description VAE - * @default null + * Control LoRA + * @description Control LoRA model to load + * @default null */ - vae?: components["schemas"]["VAEField"] | null; + control_lora?: components["schemas"]["ControlLoRAField"] | null; /** - * Tiled - * @description Processing using overlapping tiles (reduce memory consumption) - * @default false + * Positive Text Conditioning + * @description Positive conditioning tensor + * @default null */ - tiled?: boolean; + positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * Tile Size - * @description The tile size for VAE tiling in pixels (image space). If set to 0, the default tile size for the model will be used. Larger tile sizes generally produce better results at the cost of higher memory usage. - * @default 0 + * Negative Text Conditioning + * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. + * @default null */ - tile_size?: number; + negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * Fp32 - * @description Whether or not to use full float32 precision - * @default false + * Redux Conditioning + * @description FLUX Redux conditioning tensor. + * @default null */ - fp32?: boolean; + redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; /** - * Color Compensation - * @description Apply VAE scaling compensation when encoding images (reduces color drift). - * @default None - * @enum {string} + * @description FLUX Fill conditioning. + * @default null */ - color_compensation?: "None" | "SDXL"; + fill_conditioning?: components["schemas"]["FluxFillConditioningField"] | null; /** - * type - * @default i2l - * @constant + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 1 */ - type: "i2l"; - }; - /** ImageUploadEntry */ - ImageUploadEntry: { - /** @description The image DTO */ - image_dto: components["schemas"]["ImageDTO"]; + cfg_scale?: number | number[]; /** - * Presigned Url - * @description The URL to get the presigned URL for the image upload + * CFG Scale Start Step + * @description Index of the first step to apply cfg_scale. Negative indices count backwards from the the last step (e.g. a value of -1 refers to the final step). + * @default 0 */ - presigned_url: string; - }; - /** - * ImageUrlsDTO - * @description The URLs for an image and its thumbnail. - */ - ImageUrlsDTO: { + cfg_scale_start_step?: number; /** - * Image Name - * @description The unique name of the image. + * CFG Scale End Step + * @description Index of the last step to apply cfg_scale. Negative indices count backwards from the last step (e.g. a value of -1 refers to the final step). + * @default -1 */ - image_name: string; + cfg_scale_end_step?: number; /** - * Image Url - * @description The URL of the image. + * Width + * @description Width of the generated image. + * @default 1024 */ - image_url: string; + width?: number; /** - * Thumbnail Url - * @description The URL of the image's thumbnail. + * Height + * @description Height of the generated image. + * @default 1024 */ - thumbnail_url: string; - }; - /** - * Add Invisible Watermark - * @description Add an invisible watermark to an image - */ - ImageWatermarkInvocation: { + height?: number; /** - * @description The board to save the image to - * @default null + * Num Steps + * @description Number of diffusion steps. Recommended values are schnell: 4, dev: 50. + * @default 4 */ - board?: components["schemas"]["BoardField"] | null; + num_steps?: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Scheduler + * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. + * @default euler + * @enum {string} */ - metadata?: components["schemas"]["MetadataField"] | null; + scheduler?: "euler" | "heun" | "lcm"; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Guidance + * @description The guidance strength. Higher values adhere more strictly to the prompt, and will produce less diverse images. FLUX dev only, ignored for schnell. + * @default 4 */ - id: string; + guidance?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Seed + * @description Randomness seed for reproducibility. + * @default 0 */ - is_intermediate?: boolean; + seed?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Control + * @description ControlNet models. + * @default null */ - use_cache?: boolean; + control?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; /** - * @description The image to check + * @description VAE * @default null */ - image?: components["schemas"]["ImageField"] | null; + controlnet_vae?: components["schemas"]["VAEField"] | null; /** - * Text - * @description Watermark text - * @default InvokeAI + * IP-Adapter + * @description IP-Adapter to apply + * @default null */ - text?: string; + ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * type - * @default img_watermark - * @constant + * Kontext Conditioning + * @description FLUX Kontext conditioning (reference image). + * @default null */ - type: "img_watermark"; - }; - /** ImagesDownloaded */ - ImagesDownloaded: { + kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; /** - * Response - * @description The message to display to the user when images begin downloading + * Dype Preset + * @description DyPE preset for high-resolution generation. 'auto' enables automatically for resolutions > 1536px. 'area' enables automatically based on image area. '4k' uses optimized settings for 4K output. + * @default off + * @enum {string} */ - response?: string | null; + dype_preset?: "off" | "manual" | "auto" | "area" | "4k"; /** - * Bulk Download Item Name - * @description The name of the bulk download item for which events will be emitted + * Dype Scale + * @description DyPE magnitude (λs). Higher values = stronger extrapolation. Only used when dype_preset is not 'off'. + * @default null */ - bulk_download_item_name?: string | null; - }; - /** - * Solid Color Infill - * @description Infills transparent areas of an image with a solid color - */ - InfillColorInvocation: { + dype_scale?: number | null; /** - * @description The board to save the image to + * Dype Exponent + * @description DyPE decay speed (λt). Controls transition from low to high frequency detail. Only used when dype_preset is not 'off'. * @default null */ - board?: components["schemas"]["BoardField"] | null; + dype_exponent?: number | null; /** - * @description Optional metadata to be saved with the image - * @default null + * type + * @default flux_denoise_meta + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "flux_denoise_meta"; + }; + /** + * FluxFillConditioningField + * @description A FLUX Fill conditioning field. + */ + FluxFillConditioningField: { + /** @description The FLUX Fill reference image. */ + image: components["schemas"]["ImageField"]; + /** @description The FLUX Fill inpaint mask. */ + mask: components["schemas"]["TensorField"]; + }; + /** + * FLUX Fill Conditioning + * @description Prepare the FLUX Fill conditioning data. + */ + FluxFillInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13398,42 +14163,44 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The FLUX Fill reference image. * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description The color to use to infill - * @default { - * "r": 127, - * "g": 127, - * "b": 127, - * "a": 255 - * } + * @description The bool inpainting mask. Excluded regions should be set to False, included regions should be set to True. + * @default null */ - color?: components["schemas"]["ColorField"]; + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default infill_rgba + * @default flux_fill * @constant */ - type: "infill_rgba"; + type: "flux_fill"; }; /** - * PatchMatch Infill - * @description Infills transparent areas of an image using the PatchMatch algorithm + * FluxFillOutput + * @description The conditioning output of a FLUX Fill invocation. */ - InfillPatchMatchInvocation: { + FluxFillOutput: { /** - * @description The board to save the image to - * @default null + * Conditioning + * @description FLUX Redux conditioning tensor */ - board?: components["schemas"]["BoardField"] | null; + fill_cond: components["schemas"]["FluxFillConditioningField"]; /** - * @description Optional metadata to be saved with the image - * @default null + * type + * @default flux_fill_output + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "flux_fill_output"; + }; + /** + * FLUX IP-Adapter + * @description Collects FLUX IP-Adapter info to pass to other nodes. + */ + FluxIPAdapterInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13452,45 +14219,53 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The IP-Adapter image prompt(s). * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Downscale - * @description Run patchmatch on downscaled image to speedup infill - * @default 2 + * IP-Adapter Model + * @description The IP-Adapter model. + * @default null */ - downscale?: number; + ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Resample Mode - * @description The resampling mode - * @default bicubic - * @enum {string} + * Clip Vision Model + * @description CLIP Vision model to use. + * @default ViT-L + * @constant */ - resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; + clip_vision_model?: "ViT-L"; + /** + * Weight + * @description The weight given to the IP-Adapter + * @default 1 + */ + weight?: number | number[]; + /** + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) + * @default 0 + */ + begin_step_percent?: number; + /** + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; /** * type - * @default infill_patchmatch + * @default flux_ip_adapter * @constant */ - type: "infill_patchmatch"; + type: "flux_ip_adapter"; }; /** - * Tile Infill - * @description Infills transparent areas of an image with tiles of the image + * Flux Ideal Size + * @description Calculates the ideal size for generation to avoid duplication */ - InfillTileInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FluxIdealSizeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13509,114 +14284,68 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process - * @default null + * Width + * @description Target width + * @default 1024 */ - image?: components["schemas"]["ImageField"] | null; + width?: number; /** - * Tile Size - * @description The tile size (px) - * @default 32 + * Height + * @description Target height + * @default 576 */ - tile_size?: number; + height?: number; /** - * Seed - * @description The seed to use for tile generation (omit for random) - * @default 0 + * Multiplier + * @description Dimensional multiplier + * @default 1 */ - seed?: number; + multiplier?: number; /** * type - * @default infill_tile + * @default flux_ideal_size * @constant */ - type: "infill_tile"; + type: "flux_ideal_size"; }; /** - * Input - * @description The type of input a field accepts. - * - `Input.Direct`: The field must have its value provided directly, when the invocation and field are instantiated. - * - `Input.Connection`: The field must have its value provided by a connection. - * - `Input.Any`: The field may have its value provided either directly or by a connection. - * @enum {string} - */ - Input: "connection" | "direct" | "any"; - /** - * InputFieldJSONSchemaExtra - * @description Extra attributes to be added to input fields and their OpenAPI schema. Used during graph execution, - * and by the workflow editor during schema parsing and UI rendering. + * FluxIdealSizeOutput + * @description Base class for invocations that output an image */ - InputFieldJSONSchemaExtra: { - input: components["schemas"]["Input"]; - field_kind: components["schemas"]["FieldKind"]; - /** - * Orig Required - * @default true - */ - orig_required: boolean; - /** - * Default - * @default null - */ - default: unknown | null; - /** - * Orig Default - * @default null - */ - orig_default: unknown | null; - /** - * Ui Hidden - * @default false - */ - ui_hidden: boolean; - /** @default null */ - ui_type: components["schemas"]["UIType"] | null; - /** @default null */ - ui_component: components["schemas"]["UIComponent"] | null; + FluxIdealSizeOutput: { /** - * Ui Order - * @default null - */ - ui_order: number | null; - /** - * Ui Choice Labels - * @default null + * Width + * @description The ideal width of the image in pixels */ - ui_choice_labels: { - [key: string]: string; - } | null; + width: number; /** - * Ui Model Base - * @default null + * Height + * @description The ideal height of the image in pixels */ - ui_model_base: components["schemas"]["BaseModelType"][] | null; + height: number; /** - * Ui Model Type - * @default null + * type + * @default flux_ideal_size_output + * @constant */ - ui_model_type: components["schemas"]["ModelType"][] | null; + type: "flux_ideal_size_output"; + }; + /** + * FLUX Kontext Image Prep + * @description Prepares an image or images for use with FLUX Kontext. The first/single image is resized to the nearest + * preferred Kontext resolution. All other images are concatenated horizontally, maintaining their aspect ratio. + */ + FluxKontextConcatenateImagesInvocation: { /** - * Ui Model Variant + * @description The board to save the image to * @default null */ - ui_model_variant: (components["schemas"]["ClipVariantType"] | components["schemas"]["ModelVariantType"])[] | null; + board?: components["schemas"]["BoardField"] | null; /** - * Ui Model Format + * @description Optional metadata to be saved with the image * @default null */ - ui_model_format: components["schemas"]["ModelFormat"][] | null; - }; - /** - * InstallStatus - * @description State of an install job running in the background. - * @enum {string} - */ - InstallStatus: "waiting" | "downloading" | "downloads_done" | "running" | "paused" | "completed" | "error" | "cancelled"; - /** - * Integer Batch - * @description Create a batched generation, where the workflow is executed once for each integer in the batch. - */ - IntegerBatchInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13635,30 +14364,37 @@ export type components = { */ use_cache?: boolean; /** - * Batch Group - * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. - * @default None - * @enum {string} + * Images + * @description The images to concatenate + * @default null */ - batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; + images?: components["schemas"]["ImageField"][] | null; /** - * Integers - * @description The integers to batch over - * @default null + * Use Preferred Resolution + * @description Use FLUX preferred resolutions for the first image + * @default true */ - integers?: number[] | null; + use_preferred_resolution?: boolean; /** * type - * @default integer_batch + * @default flux_kontext_image_prep * @constant */ - type: "integer_batch"; + type: "flux_kontext_image_prep"; }; /** - * Integer Collection Primitive - * @description A collection of integer primitive values + * FluxKontextConditioningField + * @description A conditioning field for FLUX Kontext (reference image). */ - IntegerCollectionInvocation: { + FluxKontextConditioningField: { + /** @description The Kontext reference image. */ + image: components["schemas"]["ImageField"]; + }; + /** + * Flux Kontext Ideal Size + * @description Calculates the ideal size for generation using Flux Kontext + */ + FluxKontextIdealSizeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13677,40 +14413,29 @@ export type components = { */ use_cache?: boolean; /** - * Collection - * @description The collection of integer values - * @default [] - */ - collection?: number[]; - /** - * type - * @default integer_collection - * @constant + * Width + * @description Target width + * @default 1024 */ - type: "integer_collection"; - }; - /** - * IntegerCollectionOutput - * @description Base class for nodes that output a collection of integers - */ - IntegerCollectionOutput: { + width?: number; /** - * Collection - * @description The int collection + * Height + * @description Target height + * @default 576 */ - collection: number[]; + height?: number; /** * type - * @default integer_collection_output + * @default flux_kontext_ideal_size * @constant */ - type: "integer_collection_output"; + type: "flux_kontext_ideal_size"; }; /** - * Integer Generator - * @description Generated a range of integers for use in a batched generation + * Kontext Conditioning - FLUX + * @description Prepares a reference image for FLUX Kontext conditioning. */ - IntegerGenerator: { + FluxKontextInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13729,38 +14454,39 @@ export type components = { */ use_cache?: boolean; /** - * Generator Type - * @description The integer generator. + * @description The Kontext reference image. + * @default null */ - generator: components["schemas"]["IntegerGeneratorField"]; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default integer_generator + * @default flux_kontext * @constant */ - type: "integer_generator"; + type: "flux_kontext"; }; - /** IntegerGeneratorField */ - IntegerGeneratorField: Record; - /** IntegerGeneratorOutput */ - IntegerGeneratorOutput: { + /** + * FluxKontextOutput + * @description The conditioning output of a FLUX Kontext invocation. + */ + FluxKontextOutput: { /** - * Integers - * @description The generated integers + * Kontext Conditioning + * @description FLUX Kontext conditioning (reference image) */ - integers: number[]; + kontext_cond: components["schemas"]["FluxKontextConditioningField"]; /** * type - * @default integer_generator_output + * @default flux_kontext_output * @constant */ - type: "integer_generator_output"; + type: "flux_kontext_output"; }; /** - * Integer Primitive - * @description An integer primitive value + * Apply LoRA - FLUX + * @description Apply a LoRA model to a FLUX transformer and/or text encoder. */ - IntegerInvocation: { + FluxLoRALoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13779,23 +14505,77 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The integer value - * @default 0 + * LoRA + * @description LoRA model to load + * @default null */ - value?: number; + lora?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 + */ + weight?: number; + /** + * FLUX Transformer + * @description Transformer + * @default null + */ + transformer?: components["schemas"]["TransformerField"] | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * T5 Encoder + * @description T5 tokenizer and text encoder + * @default null + */ + t5_encoder?: components["schemas"]["T5EncoderField"] | null; /** * type - * @default integer + * @default flux_lora_loader * @constant */ - type: "integer"; + type: "flux_lora_loader"; }; /** - * Integer Math - * @description Performs integer math. + * FluxLoRALoaderOutput + * @description FLUX LoRA Loader Output */ - IntegerMathInvocation: { + FluxLoRALoaderOutput: { + /** + * FLUX Transformer + * @description Transformer + * @default null + */ + transformer: components["schemas"]["TransformerField"] | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip: components["schemas"]["CLIPField"] | null; + /** + * T5 Encoder + * @description T5 tokenizer and text encoder + * @default null + */ + t5_encoder: components["schemas"]["T5EncoderField"] | null; + /** + * type + * @default flux_lora_loader_output + * @constant + */ + type: "flux_lora_loader_output"; + }; + /** + * FLUX Main Model Input + * @description Loads a flux model from an input, outputting its submodels. + */ + FluxModelLoaderInputInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13814,53 +14594,40 @@ export type components = { */ use_cache?: boolean; /** - * Operation - * @description The operation to perform - * @default ADD - * @enum {string} - */ - operation?: "ADD" | "SUB" | "MUL" | "DIV" | "EXP" | "MOD" | "ABS" | "MIN" | "MAX"; - /** - * A - * @description The first number - * @default 1 + * @description Flux model (Transformer) to load + * @default null */ - a?: number; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * B - * @description The second number - * @default 1 + * T5 Encoder + * @description T5 tokenizer and text encoder + * @default null */ - b?: number; + t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * type - * @default integer_math - * @constant + * CLIP Embed + * @description CLIP Embed loader + * @default null */ - type: "integer_math"; - }; - /** - * IntegerOutput - * @description Base class for nodes that output a single integer - */ - IntegerOutput: { + clip_embed_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Value - * @description The output integer + * VAE + * @description VAE model to load + * @default null */ - value: number; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; /** * type - * @default integer_output + * @default flux_model_loader_input * @constant */ - type: "integer_output"; + type: "flux_model_loader_input"; }; /** - * Invert Tensor Mask - * @description Inverts a tensor mask. + * Main Model - FLUX + * @description Loads a flux base model, outputting its submodels. */ - InvertTensorMaskInvocation: { + FluxModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13879,300 +14646,6704 @@ export type components = { */ use_cache?: boolean; /** - * @description The tensor mask to convert. + * @description Flux model (Transformer) to load * @default null */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default invert_tensor_mask - * @constant - */ - type: "invert_tensor_mask"; - }; - /** InvocationCacheStatus */ - InvocationCacheStatus: { - /** - * Size - * @description The current size of the invocation cache - */ - size: number; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Hits - * @description The number of cache hits + * T5 Encoder + * @description T5 tokenizer and text encoder + * @default null */ - hits: number; + t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Misses - * @description The number of cache misses + * CLIP Embed + * @description CLIP Embed loader + * @default null */ - misses: number; + clip_embed_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Enabled - * @description Whether the invocation cache is enabled + * VAE + * @description VAE model to load + * @default null */ - enabled: boolean; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Max Size - * @description The maximum size of the invocation cache + * type + * @default flux_model_loader + * @constant */ - max_size: number; + type: "flux_model_loader"; }; /** - * InvocationCompleteEvent - * @description Event model for invocation_complete + * FluxModelLoaderOutput + * @description Flux base model loader output */ - InvocationCompleteEvent: { + FluxModelLoaderOutput: { /** - * Timestamp - * @description The timestamp of the event + * Transformer + * @description Transformer */ - timestamp: number; + transformer: components["schemas"]["TransformerField"]; /** - * Queue Id - * @description The ID of the queue + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - queue_id: string; + clip: components["schemas"]["CLIPField"]; /** - * Item Id - * @description The ID of the queue item + * T5 Encoder + * @description T5 tokenizer and text encoder */ - item_id: number; + t5_encoder: components["schemas"]["T5EncoderField"]; /** - * Batch Id - * @description The ID of the queue batch + * VAE + * @description VAE */ - batch_id: string; + vae: components["schemas"]["VAEField"]; /** - * Origin - * @description The origin of the queue item - * @default null + * Max Seq Length + * @description The max sequence length to used for the T5 encoder. (256 for schnell transformer, 512 for dev transformer) + * @enum {integer} */ - origin: string | null; + max_seq_len: 256 | 512; /** - * Destination - * @description The destination of the queue item - * @default null + * type + * @default flux_model_loader_output + * @constant */ - destination: string | null; + type: "flux_model_loader_output"; + }; + /** + * Flux Model To String + * @description Converts an Flux Model to a JSONString + */ + FluxModelToStringInvocation: { /** - * User Id - * @description The ID of the user who created the queue item - * @default system + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - user_id: string; + id: string; /** - * Session Id - * @description The ID of the session (aka graph execution state) + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - session_id: string; + is_intermediate?: boolean; /** - * Invocation - * @description The ID of the invocation + * Use Cache + * @description Whether or not to use the cache + * @default true */ - invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + use_cache?: boolean; /** - * Invocation Source Id - * @description The ID of the prepared invocation's source node + * @description Flux model (Transformer) to load + * @default null */ - invocation_source_id: string; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Result - * @description The result of the invocation + * type + * @default flux_model_to_string + * @constant */ - result: components["schemas"]["BooleanCollectionOutput"] | components["schemas"]["BooleanOutput"] | components["schemas"]["BoundingBoxCollectionOutput"] | components["schemas"]["BoundingBoxOutput"] | components["schemas"]["CLIPOutput"] | components["schemas"]["CLIPSkipInvocationOutput"] | components["schemas"]["CalculateImageTilesOutput"] | components["schemas"]["CogView4ConditioningOutput"] | components["schemas"]["CogView4ModelLoaderOutput"] | components["schemas"]["CollectInvocationOutput"] | components["schemas"]["ColorCollectionOutput"] | components["schemas"]["ColorOutput"] | components["schemas"]["ConditioningCollectionOutput"] | components["schemas"]["ConditioningOutput"] | components["schemas"]["ControlOutput"] | components["schemas"]["DenoiseMaskOutput"] | components["schemas"]["FaceMaskOutput"] | components["schemas"]["FaceOffOutput"] | components["schemas"]["FloatCollectionOutput"] | components["schemas"]["FloatGeneratorOutput"] | components["schemas"]["FloatOutput"] | components["schemas"]["Flux2KleinLoRALoaderOutput"] | components["schemas"]["Flux2KleinModelLoaderOutput"] | components["schemas"]["FluxConditioningCollectionOutput"] | components["schemas"]["FluxConditioningOutput"] | components["schemas"]["FluxControlLoRALoaderOutput"] | components["schemas"]["FluxControlNetOutput"] | components["schemas"]["FluxFillOutput"] | components["schemas"]["FluxKontextOutput"] | components["schemas"]["FluxLoRALoaderOutput"] | components["schemas"]["FluxModelLoaderOutput"] | components["schemas"]["FluxReduxOutput"] | components["schemas"]["GradientMaskOutput"] | components["schemas"]["IPAdapterOutput"] | components["schemas"]["IdealSizeOutput"] | components["schemas"]["ImageCollectionOutput"] | components["schemas"]["ImageGeneratorOutput"] | components["schemas"]["ImageOutput"] | components["schemas"]["ImagePanelCoordinateOutput"] | components["schemas"]["IntegerCollectionOutput"] | components["schemas"]["IntegerGeneratorOutput"] | components["schemas"]["IntegerOutput"] | components["schemas"]["IterateInvocationOutput"] | components["schemas"]["LatentsCollectionOutput"] | components["schemas"]["LatentsMetaOutput"] | components["schemas"]["LatentsOutput"] | components["schemas"]["LoRALoaderOutput"] | components["schemas"]["LoRASelectorOutput"] | components["schemas"]["MDControlListOutput"] | components["schemas"]["MDIPAdapterListOutput"] | components["schemas"]["MDT2IAdapterListOutput"] | components["schemas"]["MaskOutput"] | components["schemas"]["MetadataItemOutput"] | components["schemas"]["MetadataOutput"] | components["schemas"]["MetadataToLorasCollectionOutput"] | components["schemas"]["MetadataToModelOutput"] | components["schemas"]["MetadataToSDXLModelOutput"] | components["schemas"]["ModelIdentifierOutput"] | components["schemas"]["ModelLoaderOutput"] | components["schemas"]["NoiseOutput"] | components["schemas"]["PBRMapsOutput"] | components["schemas"]["PairTileImageOutput"] | components["schemas"]["PromptTemplateOutput"] | components["schemas"]["SD3ConditioningOutput"] | components["schemas"]["SDXLLoRALoaderOutput"] | components["schemas"]["SDXLModelLoaderOutput"] | components["schemas"]["SDXLRefinerModelLoaderOutput"] | components["schemas"]["SchedulerOutput"] | components["schemas"]["Sd3ModelLoaderOutput"] | components["schemas"]["SeamlessModeOutput"] | components["schemas"]["String2Output"] | components["schemas"]["StringCollectionOutput"] | components["schemas"]["StringGeneratorOutput"] | components["schemas"]["StringOutput"] | components["schemas"]["StringPosNegOutput"] | components["schemas"]["T2IAdapterOutput"] | components["schemas"]["TileToPropertiesOutput"] | components["schemas"]["UNetOutput"] | components["schemas"]["VAEOutput"] | components["schemas"]["ZImageConditioningOutput"] | components["schemas"]["ZImageControlOutput"] | components["schemas"]["ZImageLoRALoaderOutput"] | components["schemas"]["ZImageModelLoaderOutput"]; + type: "flux_model_to_string"; }; /** - * InvocationErrorEvent - * @description Event model for invocation_error + * Flux Redux Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option */ - InvocationErrorEvent: { + FluxReduxCollectionIndexInvocation: { /** - * Timestamp - * @description The timestamp of the event + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - timestamp: number; + id: string; /** - * Queue Id - * @description The ID of the queue + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - queue_id: string; + is_intermediate?: boolean; /** - * Item Id - * @description The ID of the queue item + * Use Cache + * @description Whether or not to use the cache + * @default false */ - item_id: number; + use_cache?: boolean; /** - * Batch Id - * @description The ID of the queue batch + * Random + * @description Random Index? + * @default true */ - batch_id: string; + random?: boolean; /** - * Origin - * @description The origin of the queue item + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 + */ + index?: number; + /** + * FLUX Redux Collection + * @description Conditioning tensor + * @default [] + */ + collection?: components["schemas"]["FluxReduxConditioningField"][]; + /** + * type + * @default flux_redux_index + * @constant + */ + type: "flux_redux_index"; + }; + /** + * FLUX Redux Collection Primitive + * @description A collection of flux redux primitive values + */ + FluxReduxCollectionInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * FLUX Redux Collection + * @description FLUX Redux Collection * @default null */ - origin: string | null; + collection?: components["schemas"]["FluxReduxConditioningField"][] | null; /** - * Destination - * @description The destination of the queue item + * type + * @default flux_redux_collection + * @constant + */ + type: "flux_redux_collection"; + }; + /** + * FLUX Redux Collection join + * @description Join a flux redux tensor or collections into a single collection of flux redux tensors + */ + FluxReduxCollectionJoinInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * FLUX Redux or Collection + * @description FLUX Reduxs * @default null */ - destination: string | null; + conditionings_a?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; /** - * User Id - * @description The ID of the user who created the queue item - * @default system + * FLUX Redux or Collection + * @description FLUX Reduxs + * @default null */ - user_id: string; + conditionings_b?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; /** - * Session Id - * @description The ID of the session (aka graph execution state) + * type + * @default flux_redux_collection_join + * @constant */ - session_id: string; + type: "flux_redux_collection_join"; + }; + /** FluxReduxCollectionOutput */ + FluxReduxCollectionOutput: { /** - * Invocation - * @description The ID of the invocation + * FLUX Redux Collection + * @description ControlNet(s) to apply */ - invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + collection: components["schemas"]["FluxReduxConditioningField"][]; /** - * Invocation Source Id - * @description The ID of the prepared invocation's source node + * type + * @default flux_redux_collection_output + * @constant */ - invocation_source_id: string; + type: "flux_redux_collection_output"; + }; + /** + * FluxReduxConditioningField + * @description A FLUX Redux conditioning tensor primitive value + */ + FluxReduxConditioningField: { + /** @description The Redux image conditioning tensor. */ + conditioning: components["schemas"]["TensorField"]; /** - * Error Type - * @description The error type + * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. + * @default null */ - error_type: string; + mask?: components["schemas"]["TensorField"] | null; + }; + /** + * FLUX Redux Conditioning Math + * @description Performs a Math operation on two FLUX Redux conditionings. + */ + FluxReduxConditioningMathOperationInvocation: { /** - * Error Message - * @description The error message + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - error_message: string; + id: string; /** - * Error Traceback - * @description The error traceback + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - error_traceback: string; + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description First FLUX Redux Conditioning (A) + * @default null + */ + cond_a?: components["schemas"]["FluxReduxConditioningField"] | null; + /** + * @description Second FLUX Redux Conditioning (B) + * @default null + */ + cond_b?: components["schemas"]["FluxReduxConditioningField"] | null; + /** + * Operation + * @description Operation to perform (A op B) + * @default ADD + * @enum {string} + */ + operation?: "ADD" | "SUB" | "MUL" | "DIV" | "APPEND" | "SPV" | "NSPV" | "LERP" | "SLERP"; + /** + * Scale + * @description Scaling factor + * @default 1 + */ + scale?: number; + /** + * Rescale Target Norm + * @description If > 0, rescales the output embeddings to the target max norm. Set to 0 to disable. + * @default 0 + */ + rescale_target_norm?: number; + /** + * type + * @default flux_redux_conditioning_math + * @constant + */ + type: "flux_redux_conditioning_math"; }; - InvocationOutputMap: { - add: components["schemas"]["IntegerOutput"]; - alpha_mask_to_tensor: components["schemas"]["MaskOutput"]; - apply_mask_to_image: components["schemas"]["ImageOutput"]; - apply_tensor_mask_to_image: components["schemas"]["ImageOutput"]; - blank_image: components["schemas"]["ImageOutput"]; - boolean: components["schemas"]["BooleanOutput"]; - boolean_collection: components["schemas"]["BooleanCollectionOutput"]; - bounding_box: components["schemas"]["BoundingBoxOutput"]; - calculate_image_tiles: components["schemas"]["CalculateImageTilesOutput"]; - calculate_image_tiles_even_split: components["schemas"]["CalculateImageTilesOutput"]; - calculate_image_tiles_min_overlap: components["schemas"]["CalculateImageTilesOutput"]; - canny_edge_detection: components["schemas"]["ImageOutput"]; - canvas_paste_back: components["schemas"]["ImageOutput"]; - canvas_v2_mask_and_crop: components["schemas"]["ImageOutput"]; - clip_skip: components["schemas"]["CLIPSkipInvocationOutput"]; - cogview4_denoise: components["schemas"]["LatentsOutput"]; - cogview4_i2l: components["schemas"]["LatentsOutput"]; - cogview4_l2i: components["schemas"]["ImageOutput"]; - cogview4_model_loader: components["schemas"]["CogView4ModelLoaderOutput"]; - cogview4_text_encoder: components["schemas"]["CogView4ConditioningOutput"]; - collect: components["schemas"]["CollectInvocationOutput"]; - color: components["schemas"]["ColorOutput"]; - color_correct: components["schemas"]["ImageOutput"]; - color_map: components["schemas"]["ImageOutput"]; - compel: components["schemas"]["ConditioningOutput"]; - conditioning: components["schemas"]["ConditioningOutput"]; - conditioning_collection: components["schemas"]["ConditioningCollectionOutput"]; - content_shuffle: components["schemas"]["ImageOutput"]; - controlnet: components["schemas"]["ControlOutput"]; - core_metadata: components["schemas"]["MetadataOutput"]; - create_denoise_mask: components["schemas"]["DenoiseMaskOutput"]; - create_gradient_mask: components["schemas"]["GradientMaskOutput"]; - crop_image_to_bounding_box: components["schemas"]["ImageOutput"]; - crop_latents: components["schemas"]["LatentsOutput"]; - cv_inpaint: components["schemas"]["ImageOutput"]; - decode_watermark: components["schemas"]["StringOutput"]; - denoise_latents: components["schemas"]["LatentsOutput"]; - denoise_latents_meta: components["schemas"]["LatentsMetaOutput"]; - depth_anything_depth_estimation: components["schemas"]["ImageOutput"]; - div: components["schemas"]["IntegerOutput"]; - dw_openpose_detection: components["schemas"]["ImageOutput"]; - dynamic_prompt: components["schemas"]["StringCollectionOutput"]; - esrgan: components["schemas"]["ImageOutput"]; - expand_mask_with_fade: components["schemas"]["ImageOutput"]; - face_identifier: components["schemas"]["ImageOutput"]; - face_mask_detection: components["schemas"]["FaceMaskOutput"]; - face_off: components["schemas"]["FaceOffOutput"]; - float: components["schemas"]["FloatOutput"]; - float_batch: components["schemas"]["FloatOutput"]; - float_collection: components["schemas"]["FloatCollectionOutput"]; - float_generator: components["schemas"]["FloatGeneratorOutput"]; - float_math: components["schemas"]["FloatOutput"]; - float_range: components["schemas"]["FloatCollectionOutput"]; - float_to_int: components["schemas"]["IntegerOutput"]; - flux2_denoise: components["schemas"]["LatentsOutput"]; - flux2_klein_lora_collection_loader: components["schemas"]["Flux2KleinLoRALoaderOutput"]; - flux2_klein_lora_loader: components["schemas"]["Flux2KleinLoRALoaderOutput"]; - flux2_klein_model_loader: components["schemas"]["Flux2KleinModelLoaderOutput"]; - flux2_klein_text_encoder: components["schemas"]["FluxConditioningOutput"]; - flux2_vae_decode: components["schemas"]["ImageOutput"]; - flux2_vae_encode: components["schemas"]["LatentsOutput"]; - flux_control_lora_loader: components["schemas"]["FluxControlLoRALoaderOutput"]; - flux_controlnet: components["schemas"]["FluxControlNetOutput"]; - flux_denoise: components["schemas"]["LatentsOutput"]; - flux_denoise_meta: components["schemas"]["LatentsMetaOutput"]; - flux_fill: components["schemas"]["FluxFillOutput"]; - flux_ip_adapter: components["schemas"]["IPAdapterOutput"]; - flux_kontext: components["schemas"]["FluxKontextOutput"]; - flux_kontext_image_prep: components["schemas"]["ImageOutput"]; - flux_lora_collection_loader: components["schemas"]["FluxLoRALoaderOutput"]; - flux_lora_loader: components["schemas"]["FluxLoRALoaderOutput"]; - flux_model_loader: components["schemas"]["FluxModelLoaderOutput"]; - flux_redux: components["schemas"]["FluxReduxOutput"]; - flux_text_encoder: components["schemas"]["FluxConditioningOutput"]; - flux_vae_decode: components["schemas"]["ImageOutput"]; - flux_vae_encode: components["schemas"]["LatentsOutput"]; - freeu: components["schemas"]["UNetOutput"]; - get_image_mask_bounding_box: components["schemas"]["BoundingBoxOutput"]; - grounding_dino: components["schemas"]["BoundingBoxCollectionOutput"]; - hed_edge_detection: components["schemas"]["ImageOutput"]; - heuristic_resize: components["schemas"]["ImageOutput"]; - i2l: components["schemas"]["LatentsOutput"]; - ideal_size: components["schemas"]["IdealSizeOutput"]; - image: components["schemas"]["ImageOutput"]; - image_batch: components["schemas"]["ImageOutput"]; - image_collection: components["schemas"]["ImageCollectionOutput"]; - image_generator: components["schemas"]["ImageGeneratorOutput"]; - image_mask_to_tensor: components["schemas"]["MaskOutput"]; - image_panel_layout: components["schemas"]["ImagePanelCoordinateOutput"]; - img_blur: components["schemas"]["ImageOutput"]; - img_chan: components["schemas"]["ImageOutput"]; - img_channel_multiply: components["schemas"]["ImageOutput"]; - img_channel_offset: components["schemas"]["ImageOutput"]; - img_conv: components["schemas"]["ImageOutput"]; - img_crop: components["schemas"]["ImageOutput"]; - img_hue_adjust: components["schemas"]["ImageOutput"]; - img_ilerp: components["schemas"]["ImageOutput"]; - img_lerp: components["schemas"]["ImageOutput"]; - img_mul: components["schemas"]["ImageOutput"]; - img_noise: components["schemas"]["ImageOutput"]; - img_nsfw: components["schemas"]["ImageOutput"]; + /** + * FLUX Redux Downsampling + * @description Downsampling Flux Redux conditioning + */ + FluxReduxDownsamplingInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description FLUX Redux conditioning tensor. + * @default null + */ + redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | null; + /** + * Downsampling Factor + * @description Redux Downsampling Factor (1-9) + * @default 3 + */ + downsampling_factor?: number; + /** + * Downsampling Function + * @description Redux Downsampling Function + * @default area + * @enum {string} + */ + downsampling_function?: "nearest" | "bilinear" | "bicubic" | "area" | "nearest-exact"; + /** + * Weight + * @description Redux weight (0.0-3.0) + * @default 1 + */ + weight?: number; + /** + * type + * @default flux_redux_downsampling + * @constant + */ + type: "flux_redux_downsampling"; + }; + /** + * FLUX Redux + * @description Runs a FLUX Redux model to generate a conditioning tensor. + */ + FluxReduxInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The FLUX Redux image prompt. + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description The bool mask associated with this FLUX Redux image prompt. Excluded regions should be set to False, included regions should be set to True. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * FLUX Redux Model + * @description The FLUX Redux model to use. + * @default null + */ + redux_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Downsampling Factor + * @description Redux Downsampling Factor (1-9) + * @default 1 + */ + downsampling_factor?: number; + /** + * Downsampling Function + * @description Redux Downsampling Function + * @default area + * @enum {string} + */ + downsampling_function?: "nearest" | "bilinear" | "bicubic" | "area" | "nearest-exact"; + /** + * Weight + * @description Redux weight (0.0-1.0) + * @default 1 + */ + weight?: number; + /** + * type + * @default flux_redux + * @constant + */ + type: "flux_redux"; + }; + /** + * FLUX Redux-Linked + * @description Collects FLUX Redux conditioning info to pass to other nodes. + */ + FluxReduxLinkedInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The FLUX Redux image prompt. + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description The bool mask associated with this FLUX Redux image prompt. Excluded regions should be set to False, included regions should be set to True. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * FLUX Redux Model + * @description The FLUX Redux model to use. + * @default null + */ + redux_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Downsampling Factor + * @description Redux Downsampling Factor (1-9) + * @default 1 + */ + downsampling_factor?: number; + /** + * Downsampling Function + * @description Redux Downsampling Function + * @default area + * @enum {string} + */ + downsampling_function?: "nearest" | "bilinear" | "bicubic" | "area" | "nearest-exact"; + /** + * Weight + * @description Redux weight (0.0-1.0) + * @default 1 + */ + weight?: number; + /** + * type + * @default flux_redux_linked + * @constant + */ + type: "flux_redux_linked"; + /** + * FLUX Redux Conditioning List + * @description FLUX Redux conditioning list + * @default null + */ + redux_conditioning_list?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; + }; + /** FluxReduxListOutput */ + FluxReduxListOutput: { + /** + * FLUX Redux Conditioning List + * @description FLUX Redux conditioning tensor + */ + redux_conditioning_list: components["schemas"]["FluxReduxConditioningField"][]; + /** + * type + * @default flux_redux_list_output + * @constant + */ + type: "flux_redux_list_output"; + }; + /** + * FluxReduxOutput + * @description The conditioning output of a FLUX Redux invocation. + */ + FluxReduxOutput: { + /** + * Conditioning + * @description FLUX Redux conditioning tensor + */ + redux_cond: components["schemas"]["FluxReduxConditioningField"]; + /** + * type + * @default flux_redux_output + * @constant + */ + type: "flux_redux_output"; + }; + /** + * Rescale FLUX Redux Conditioning + * @description Rescales a FLUX Redux conditioning field to a target max norm. + */ + FluxReduxRescaleConditioningInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description FLUX Redux Conditioning to rescale + * @default null + */ + redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | null; + /** + * Target Norm + * @description The target max norm for the conditioning. + * @default 1 + */ + target_norm?: number; + /** + * type + * @default flux_redux_rescale_conditioning + * @constant + */ + type: "flux_redux_rescale_conditioning"; + }; + /** + * Rescale FLUX Conditioning + * @description Rescales a FLUX conditioning field to a target max norm. + */ + FluxRescaleConditioningInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description FLUX Conditioning to rescale + * @default null + */ + conditioning?: components["schemas"]["FluxConditioningField"] | null; + /** + * Clip Rescale + * @description Whether to rescale the CLIP embeddings. + * @default true + */ + clip_rescale?: boolean; + /** + * Clip Target Norm + * @description The target max norm for the CLIP embeddings. + * @default 30 + */ + clip_target_norm?: number; + /** + * T5 Rescale + * @description Whether to rescale the T5 embeddings. + * @default true + */ + t5_rescale?: boolean; + /** + * T5 Target Norm + * @description The target max norm for the T5 embeddings. + * @default 10 + */ + t5_target_norm?: number; + /** + * type + * @default flux_rescale_conditioning + * @constant + */ + type: "flux_rescale_conditioning"; + }; + /** + * Scale FLUX Conditioning + * @description Scales a FLUX conditioning field by a factor. + */ + FluxScaleConditioningInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description FLUX Conditioning to scale + * @default null + */ + conditioning?: components["schemas"]["FluxConditioningField"] | null; + /** + * Scale + * @description Scaling factor + * @default 1 + */ + scale?: number; + /** + * Negative + * @description Scale negative conditioning + * @default false + */ + negative?: boolean; + /** + * type + * @default flux_scale_conditioning + * @constant + */ + type: "flux_scale_conditioning"; + }; + /** + * Scale FLUX Prompt Section(s) + * @description Scales one or more sections of a FLUX prompt conditioning. + */ + FluxScalePromptSectionInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description FLUX Conditioning to modify + * @default null + */ + conditioning?: components["schemas"]["FluxConditioningField"] | null; + /** + * T5Encoder + * @description T5 Encoder model and tokenizer used for the original conditioning. + * @default null + */ + t5_encoder?: components["schemas"]["T5EncoderField"] | null; + /** + * Prompt + * @description The full prompt text used for the original conditioning. + * @default null + */ + prompt?: string | null; + /** + * Prompt Section + * @description The section or sections of the prompt to scale. + * @default null + */ + prompt_section?: string | string[] | null; + /** + * Scale + * @description The scaling factor or factors for the prompt section(s). + * @default 1 + */ + scale?: number | number[]; + /** + * Positions + * @description The start character position(s) of the section(s) to scale. If provided, this is used to locate the section(s) instead of searching. + * @default null + */ + positions?: number | number[] | null; + /** + * Rescale Output + * @description Rescales the output T5 embeddings to have the same max vector norm as the original conditioning. + * @default false + */ + rescale_output?: boolean; + /** + * type + * @default flux_scale_prompt_section + * @constant + */ + type: "flux_scale_prompt_section"; + }; + /** + * Scale FLUX Redux Conditioning + * @description Scales a FLUX Redux conditioning field by a factor. + */ + FluxScaleReduxConditioningInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description FLUX Redux Conditioning to scale + * @default null + */ + redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | null; + /** + * Scale + * @description Scaling factor + * @default 1 + */ + scale?: number; + /** + * Negative + * @description Scale negative conditioning + * @default false + */ + negative?: boolean; + /** + * type + * @default flux_scale_redux_conditioning + * @constant + */ + type: "flux_scale_redux_conditioning"; + }; + /** + * Prompt - FLUX + * @description Encodes and preps a prompt for a flux image. + */ + FluxTextEncoderInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * T5Encoder + * @description T5 tokenizer and text encoder + * @default null + */ + t5_encoder?: components["schemas"]["T5EncoderField"] | null; + /** + * T5 Max Seq Len + * @description Max sequence length for the T5 encoder. Expected to be 256 for FLUX schnell models and 512 for FLUX dev models. + * @default null + */ + t5_max_seq_len?: (256 | 512) | null; + /** + * Prompt + * @description Text prompt to encode. + * @default null + */ + prompt?: string | null; + /** + * @description A mask defining the region that this conditioning prompt applies to. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * type + * @default flux_text_encoder + * @constant + */ + type: "flux_text_encoder"; + }; + /** + * Prompt - FLUX Linked + * @description Collects FLUX prompt conditionings to pass to other nodes. + */ + FluxTextEncoderLinkedInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * T5Encoder + * @description T5 tokenizer and text encoder + * @default null + */ + t5_encoder?: components["schemas"]["T5EncoderField"] | null; + /** + * T5 Max Seq Len + * @description Max sequence length for the T5 encoder. Expected to be 256 for FLUX schnell models and 512 for FLUX dev models. + * @default null + */ + t5_max_seq_len?: (256 | 512) | null; + /** + * Prompt + * @description Text prompt to encode. + * @default null + */ + prompt?: string | null; + /** + * @description A mask defining the region that this conditioning prompt applies to. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * type + * @default flux_text_encoder_linked + * @constant + */ + type: "flux_text_encoder_linked"; + /** + * FLUX Text Encoder Conditioning List + * @description Conditioning tensor + * @default null + */ + flux_text_encoder_list?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + }; + /** + * FluxTextEncoderListOutput + * @description Output for a list of FLUX text encoder conditioning. + */ + FluxTextEncoderListOutput: { + /** + * FLUX Text Encoder Conditioning List + * @description Conditioning tensor + */ + flux_text_encoder_list: components["schemas"]["FluxConditioningField"][]; + /** + * type + * @default flux_text_encoder_list_output + * @constant + */ + type: "flux_text_encoder_list_output"; + }; + /** + * Latents to Image - FLUX + * @description Generates an image from latents. + */ + FluxVaeDecodeInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latents?: components["schemas"]["LatentsField"] | null; + /** + * @description VAE + * @default null + */ + vae?: components["schemas"]["VAEField"] | null; + /** + * type + * @default flux_vae_decode + * @constant + */ + type: "flux_vae_decode"; + }; + /** + * Image to Latents - FLUX + * @description Encodes an image into latents. + */ + FluxVaeEncodeInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to encode. + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description VAE + * @default null + */ + vae?: components["schemas"]["VAEField"] | null; + /** + * type + * @default flux_vae_encode + * @constant + */ + type: "flux_vae_encode"; + }; + /** + * FluxVariantType + * @description FLUX.1 model variants. + * @enum {string} + */ + FluxVariantType: "schnell" | "dev" | "dev_fill"; + /** + * FLUX Weighted Prompt + * @description Parses a weighted prompt, then encodes it for FLUX. + */ + FluxWeightedPromptInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Prompt + * @description Text prompt to encode. + * @default null + */ + prompt?: string | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * T5Encoder + * @description T5 tokenizer and text encoder + * @default null + */ + t5_encoder?: components["schemas"]["T5EncoderField"] | null; + /** + * T5 Max Seq Len + * @description Max sequence length for the T5 encoder. Expected to be 256 for FLUX schnell models and 512 for FLUX dev models. + * @default null + */ + t5_max_seq_len?: (256 | 512) | null; + /** + * @description A mask defining the region that this conditioning prompt applies to. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * Rescale Output + * @description Rescales the output T5 embeddings to have the same max vector norm as the original conditioning. + * @default false + */ + rescale_output?: boolean; + /** + * type + * @default flux_weighted_prompt + * @constant + */ + type: "flux_weighted_prompt"; + }; + /** + * FluxWeightedPromptOutput + * @description Outputs a FLUX conditioning and a cleaned prompt + */ + FluxWeightedPromptOutput: { + /** @description The FLUX conditioning */ + conditioning: components["schemas"]["FluxConditioningField"]; + /** + * Cleaned Prompt + * @description The prompt with all weighting syntax removed + */ + cleaned_prompt: string; + /** + * type + * @default flux_weighted_prompt_output + * @constant + */ + type: "flux_weighted_prompt_output"; + }; + /** FoundModel */ + FoundModel: { + /** + * Path + * @description Path to the model + */ + path: string; + /** + * Is Installed + * @description Whether or not the model is already installed + */ + is_installed: boolean; + }; + /** + * FreeUConfig + * @description Configuration for the FreeU hyperparameters. + * - https://huggingface.co/docs/diffusers/main/en/using-diffusers/freeu + * - https://github.com/ChenyangSi/FreeU + */ + FreeUConfig: { + /** + * S1 + * @description Scaling factor for stage 1 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. + */ + s1: number; + /** + * S2 + * @description Scaling factor for stage 2 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. + */ + s2: number; + /** + * B1 + * @description Scaling factor for stage 1 to amplify the contributions of backbone features. + */ + b1: number; + /** + * B2 + * @description Scaling factor for stage 2 to amplify the contributions of backbone features. + */ + b2: number; + }; + /** + * Apply FreeU - SD1.5, SDXL + * @description Applies FreeU to the UNet. Suggested values (b1/b2/s1/s2): + * + * SD1.5: 1.2/1.4/0.9/0.2, + * SD2: 1.1/1.2/0.9/0.2, + * SDXL: 1.1/1.2/0.6/0.4, + */ + FreeUInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * UNet + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet?: components["schemas"]["UNetField"] | null; + /** + * B1 + * @description Scaling factor for stage 1 to amplify the contributions of backbone features. + * @default 1.2 + */ + b1?: number; + /** + * B2 + * @description Scaling factor for stage 2 to amplify the contributions of backbone features. + * @default 1.4 + */ + b2?: number; + /** + * S1 + * @description Scaling factor for stage 1 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. + * @default 0.9 + */ + s1?: number; + /** + * S2 + * @description Scaling factor for stage 2 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. + * @default 0.2 + */ + s2?: number; + /** + * type + * @default freeu + * @constant + */ + type: "freeu"; + }; + /** Frequency Blend Latents */ + FrequencyBlendLatents: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Image from which to take high frequencies. + * @default null + */ + high_input?: components["schemas"]["LatentsField"] | null; + /** + * @description Image from which to take low frequencies. + * @default null + */ + low_input?: components["schemas"]["LatentsField"] | null; + /** + * Threshold + * @description How much of the high-frequency to use. Negative values swap the inputs. + * @default 0.05 + */ + threshold?: number; + /** + * Weighted Blend + * @description Weighted blend? + * @default true + */ + weighted_blend?: boolean; + /** + * type + * @default frequency_blend_latents + * @constant + */ + type: "frequency_blend_latents"; + }; + /** + * Frequency Spectrum Match Latents + * @description Generates latent noise with the frequency spectrum of target latents, masked by a provided mask image. + * Takes both target latents and white noise latents as inputs. + */ + FrequencySpectrumMatchLatentsInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Target latents to match frequency spectrum (A) + * @default null + */ + target_latents_a?: components["schemas"]["LatentsField"] | null; + /** + * @description Target latents to match frequency spectrum (B) (optional) + * @default null + */ + target_latents_b?: components["schemas"]["LatentsField"] | null; + /** + * Frequency Blend Alpha + * @description Blend ratio for the frequency spectra + * @default 0 + */ + frequency_blend_alpha?: number; + /** + * @description White noise latents for phase information (A) + * @default null + */ + phase_latents_a?: components["schemas"]["LatentsField"] | null; + /** + * @description White noise latents for phase information (B) (optional) + * @default null + */ + phase_latents_b?: components["schemas"]["LatentsField"] | null; + /** + * Phase Blend Alpha + * @description Blend ratio for the phases + * @default 0 + */ + phase_blend_alpha?: number; + /** + * @description Mask for blending (optional) + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * Blur Sigma + * @description Amount of Gaussian blur to apply to the mask + * @default 0 + */ + blur_sigma?: number; + /** + * type + * @default frequency_match_latents + * @constant + */ + type: "frequency_match_latents"; + }; + /** + * Generate Evolutionary Prompts + * @description Generates a new population of prompts by performing crossover operations + * on an input list of parent seed vectors. + */ + GenerateEvolutionaryPromptsInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * Resolutions Dict + * @description Private field for id substitutions dict cache + * @default {} + */ + resolutions_dict?: { + [key: string]: unknown; + }; + /** + * Lookups + * @description Lookup table(s) containing template(s) (JSON) + * @default [] + */ + lookups?: string | string[]; + /** + * Remove Negatives + * @description Whether to strip out text between [] + * @default false + */ + remove_negatives?: boolean; + /** + * Strip Parens Probability + * @description Probability of removing attention group weightings + * @default 0 + */ + strip_parens_probability?: number; + /** + * Resolutions + * @description JSON structure of substitutions by id by tag + * @default [] + */ + resolutions?: string | string[]; + /** + * Seed Vectors In + * @description List of JSON array strings, each representing a parent's seed vector. + * @default [] + */ + seed_vectors_in?: string[]; + /** + * Target Population Size + * @description The desired size of the new population to generate. + * @default 10 + */ + target_population_size?: number; + /** + * Selected Pair + * @description The selected population member to output specifically. + * @default 0 + */ + selected_pair?: number; + /** + * Ga Seed + * @description Seed for the random number generator to ensure deterministic GA operations + * @default 0 + */ + ga_seed?: number; + /** + * Crossover Non Terminal + * @description Optional: The non-terminal (key in lookups) to target for the crossover branch. If None, a random one will be chosen from available common non-terminals for each crossover. + * @default null + */ + crossover_non_terminal?: string | null; + /** + * type + * @default generate_evolutionary_prompts + * @constant + */ + type: "generate_evolutionary_prompts"; + }; + /** + * GeneratePasswordResponse + * @description Response containing a generated password. + */ + GeneratePasswordResponse: { + /** + * Password + * @description Generated strong password + */ + password: string; + }; + /** + * Get Image Mask Bounding Box + * @description Gets the bounding box of the given mask image. + */ + GetMaskBoundingBoxInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The mask to crop. + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * Margin + * @description Margin to add to the bounding box. + * @default 0 + */ + margin?: number; + /** + * @description Color of the mask in the image. + * @default { + * "r": 255, + * "g": 255, + * "b": 255, + * "a": 255 + * } + */ + mask_color?: components["schemas"]["ColorField"]; + /** + * type + * @default get_image_mask_bounding_box + * @constant + */ + type: "get_image_mask_bounding_box"; + }; + /** + * Get Source Frame Rate + * @description Get the source framerate of an MP4 video and provide it as output. + */ + GetSourceFrameRateInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Video Path + * @description The path to the MP4 video file + * @default null + */ + video_path?: string | null; + /** + * type + * @default get_source_framerate + * @constant + */ + type: "get_source_framerate"; + }; + /** + * Get Total Frames + * @description Get the total number of frames in an MP4 video and provide it as output. + */ + GetTotalFramesInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Video Path + * @description The path to the MP4 video file + * @default null + */ + video_path?: string | null; + /** + * type + * @default get_total_frames + * @constant + */ + type: "get_total_frames"; + }; + /** GlmEncoderField */ + GlmEncoderField: { + /** @description Info to load tokenizer submodel */ + tokenizer: components["schemas"]["ModelIdentifierField"]; + /** @description Info to load text_encoder submodel */ + text_encoder: components["schemas"]["ModelIdentifierField"]; + }; + /** + * GradientMaskOutput + * @description Outputs a denoise mask and an image representing the total gradient of the mask. + */ + GradientMaskOutput: { + /** @description Mask for denoise model run. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. */ + denoise_mask: components["schemas"]["DenoiseMaskField"]; + /** @description Image representing the total gradient area of the mask. For paste-back purposes. */ + expanded_mask_area: components["schemas"]["ImageField"]; + /** + * type + * @default gradient_mask_output + * @constant + */ + type: "gradient_mask_output"; + }; + /** Graph */ + Graph: { + /** + * Id + * @description The id of this graph + */ + id?: string; + /** + * Nodes + * @description The nodes in this graph + */ + nodes?: { + [key: string]: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + }; + /** + * Edges + * @description The connections between nodes and their fields in this graph + */ + edges?: components["schemas"]["Edge"][]; + }; + /** + * GraphExecutionState + * @description Tracks the state of a graph execution + */ + GraphExecutionState: { + /** + * Id + * @description The id of the execution state + */ + id: string; + /** @description The graph being executed */ + graph: components["schemas"]["Graph"]; + /** @description The expanded graph of activated and executed nodes */ + execution_graph: components["schemas"]["Graph"]; + /** + * Executed + * @description The set of node ids that have been executed + */ + executed: string[]; + /** + * Executed History + * @description The list of node ids that have been executed, in order of execution + */ + executed_history: string[]; + /** + * Results + * @description The results of node executions + */ + results: { + [key: string]: components["schemas"]["BooleanCollectionOutput"] | components["schemas"]["BooleanOutput"] | components["schemas"]["BoundingBoxCollectionOutput"] | components["schemas"]["BoundingBoxOutput"] | components["schemas"]["CLIPOutput"] | components["schemas"]["CLIPSkipInvocationOutput"] | components["schemas"]["CMYKSeparationOutput"] | components["schemas"]["CMYKSplitOutput"] | components["schemas"]["CalculateImageTilesOutput"] | components["schemas"]["ClipsegMaskHierarchyOutput"] | components["schemas"]["CogView4ConditioningOutput"] | components["schemas"]["CogView4ModelLoaderOutput"] | components["schemas"]["CollectInvocationOutput"] | components["schemas"]["CollectionCountOutput"] | components["schemas"]["CollectionIndexOutput"] | components["schemas"]["CollectionJoinOutput"] | components["schemas"]["CollectionReverseOutput"] | components["schemas"]["CollectionSliceOutput"] | components["schemas"]["CollectionSortOutput"] | components["schemas"]["CollectionUniqueOutput"] | components["schemas"]["ColorCollectionOutput"] | components["schemas"]["ColorOutput"] | components["schemas"]["ConditioningCollectionOutput"] | components["schemas"]["ConditioningOutput"] | components["schemas"]["ControlListOutput"] | components["schemas"]["ControlOutput"] | components["schemas"]["DenoiseMaskOutput"] | components["schemas"]["EscapedOutput"] | components["schemas"]["EvolutionaryPromptListOutput"] | components["schemas"]["FaceMaskOutput"] | components["schemas"]["FaceOffOutput"] | components["schemas"]["FloatCollectionOutput"] | components["schemas"]["FloatGeneratorOutput"] | components["schemas"]["FloatOutput"] | components["schemas"]["Flux2KleinLoRALoaderOutput"] | components["schemas"]["Flux2KleinModelLoaderOutput"] | components["schemas"]["FluxConditioningBlendOutput"] | components["schemas"]["FluxConditioningCollectionOutput"] | components["schemas"]["FluxConditioningDeltaAndAugmentedOutput"] | components["schemas"]["FluxConditioningListOutput"] | components["schemas"]["FluxConditioningOutput"] | components["schemas"]["FluxConditioningStoreOutput"] | components["schemas"]["FluxControlLoRALoaderOutput"] | components["schemas"]["FluxControlNetCollectionOutput"] | components["schemas"]["FluxControlNetListOutput"] | components["schemas"]["FluxControlNetOutput"] | components["schemas"]["FluxFillOutput"] | components["schemas"]["FluxIdealSizeOutput"] | components["schemas"]["FluxKontextOutput"] | components["schemas"]["FluxLoRALoaderOutput"] | components["schemas"]["FluxModelLoaderOutput"] | components["schemas"]["FluxReduxCollectionOutput"] | components["schemas"]["FluxReduxListOutput"] | components["schemas"]["FluxReduxOutput"] | components["schemas"]["FluxTextEncoderListOutput"] | components["schemas"]["FluxWeightedPromptOutput"] | components["schemas"]["GradientMaskOutput"] | components["schemas"]["HalvedPromptOutput"] | components["schemas"]["IPAdapterListOutput"] | components["schemas"]["IPAdapterOutput"] | components["schemas"]["IdealSizeOutput"] | components["schemas"]["ImageCollectionOutput"] | components["schemas"]["ImageGeneratorOutput"] | components["schemas"]["ImageIndexCollectOutput"] | components["schemas"]["ImageOutput"] | components["schemas"]["ImagePanelCoordinateOutput"] | components["schemas"]["ImageSOMOutput"] | components["schemas"]["ImageToImageNameOutput"] | components["schemas"]["ImageToXYImageTilesOutput"] | components["schemas"]["ImagesIndexToVideoOutput"] | components["schemas"]["IntegerCollectionOutput"] | components["schemas"]["IntegerGeneratorOutput"] | components["schemas"]["IntegerOutput"] | components["schemas"]["IterateInvocationOutput"] | components["schemas"]["JsonListStringsOutput"] | components["schemas"]["LatentsCollectionOutput"] | components["schemas"]["LatentsMetaOutput"] | components["schemas"]["LatentsOutput"] | components["schemas"]["LoRACollectionFromPathOutput"] | components["schemas"]["LoRACollectionOutput"] | components["schemas"]["LoRACollectionToggleOutput"] | components["schemas"]["LoRALoaderOutput"] | components["schemas"]["LoRANameGrabberOutput"] | components["schemas"]["LoRASelectorOutput"] | components["schemas"]["LoRAToggleOutput"] | components["schemas"]["LoadAllTextFilesInFolderOutput"] | components["schemas"]["LoadTextFileToStringOutput"] | components["schemas"]["LookupLoRACollectionTriggersOutput"] | components["schemas"]["LookupLoRATriggersOutput"] | components["schemas"]["LookupTableOutput"] | components["schemas"]["MDControlListOutput"] | components["schemas"]["MDIPAdapterListOutput"] | components["schemas"]["MDT2IAdapterListOutput"] | components["schemas"]["MaskOutput"] | components["schemas"]["MathEvalOutput"] | components["schemas"]["MergeLoRACollectionsOutput"] | components["schemas"]["MergeStringCollectionsOutput"] | components["schemas"]["MetadataItemOutput"] | components["schemas"]["MetadataOutput"] | components["schemas"]["MetadataToLorasCollectionOutput"] | components["schemas"]["MetadataToModelOutput"] | components["schemas"]["MetadataToSDXLModelOutput"] | components["schemas"]["ModelIdentifierOutput"] | components["schemas"]["ModelLoaderOutput"] | components["schemas"]["ModelNameGrabberOutput"] | components["schemas"]["ModelToggleOutput"] | components["schemas"]["NightmareOutput"] | components["schemas"]["NoiseOutput"] | components["schemas"]["PBRMapsOutput"] | components["schemas"]["PTFieldsCollectOutput"] | components["schemas"]["PTFieldsExpandOutput"] | components["schemas"]["PairTileImageOutput"] | components["schemas"]["PaletteOutput"] | components["schemas"]["PrintStringToConsoleOutput"] | components["schemas"]["PromptStrengthOutput"] | components["schemas"]["PromptTemplateOutput"] | components["schemas"]["PromptsToFileInvocationOutput"] | components["schemas"]["PrunedPromptOutput"] | components["schemas"]["Qwen3PromptProOutput"] | components["schemas"]["RGBSplitOutput"] | components["schemas"]["RandomImageSizeOutput"] | components["schemas"]["RandomLoRAMixerOutput"] | components["schemas"]["ReapplyLoRAWeightOutput"] | components["schemas"]["RetrieveFluxConditioningMultiOutput"] | components["schemas"]["SD3ConditioningOutput"] | components["schemas"]["SDXLLoRALoaderOutput"] | components["schemas"]["SDXLModelLoaderOutput"] | components["schemas"]["SDXLRefinerModelLoaderOutput"] | components["schemas"]["SchedulerOutput"] | components["schemas"]["Sd3ModelLoaderOutput"] | components["schemas"]["SeamlessModeOutput"] | components["schemas"]["ShadowsHighlightsMidtonesMasksOutput"] | components["schemas"]["String2Output"] | components["schemas"]["StringCollectionJoinerOutput"] | components["schemas"]["StringCollectionOutput"] | components["schemas"]["StringGeneratorOutput"] | components["schemas"]["StringOutput"] | components["schemas"]["StringPosNegOutput"] | components["schemas"]["StringToLoraOutput"] | components["schemas"]["StringToMainModelOutput"] | components["schemas"]["StringToModelOutput"] | components["schemas"]["StringToSDXLModelOutput"] | components["schemas"]["T2IAdapterListOutput"] | components["schemas"]["T2IAdapterOutput"] | components["schemas"]["ThresholdingOutput"] | components["schemas"]["TileSizeOutput"] | components["schemas"]["TileToPropertiesOutput"] | components["schemas"]["TilesOutput"] | components["schemas"]["TraceryOutput"] | components["schemas"]["UNetOutput"] | components["schemas"]["VAEOutput"] | components["schemas"]["WeightedStringOutput"] | components["schemas"]["XYExpandOutput"] | components["schemas"]["XYImageExpandOutput"] | components["schemas"]["XYProductOutput"] | components["schemas"]["ZImageConditioningOutput"] | components["schemas"]["ZImageControlOutput"] | components["schemas"]["ZImageLoRALoaderOutput"] | components["schemas"]["ZImageModelLoaderOutput"]; + }; + /** + * Errors + * @description Errors raised when executing nodes + */ + errors: { + [key: string]: string; + }; + /** + * Prepared Source Mapping + * @description The map of prepared nodes to original graph nodes + */ + prepared_source_mapping: { + [key: string]: string; + }; + /** + * Source Prepared Mapping + * @description The map of original graph nodes to prepared nodes + */ + source_prepared_mapping: { + [key: string]: string[]; + }; + /** Ready Order */ + ready_order?: string[]; + /** + * Indegree + * @description Remaining unmet input count for exec nodes + */ + indegree?: { + [key: string]: number; + }; + }; + /** + * Grounding DINO (Text Prompt Object Detection) + * @description Runs a Grounding DINO model. Performs zero-shot bounding-box object detection from a text prompt. + */ + GroundingDinoInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Model + * @description The Grounding DINO model to use. + * @default null + */ + model?: ("grounding-dino-tiny" | "grounding-dino-base") | null; + /** + * Prompt + * @description The prompt describing the object to segment. + * @default null + */ + prompt?: string | null; + /** + * @description The image to segment. + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Detection Threshold + * @description The detection threshold for the Grounding DINO model. All detected bounding boxes with scores above this threshold will be returned. + * @default 0.3 + */ + detection_threshold?: number; + /** + * type + * @default grounding_dino + * @constant + */ + type: "grounding_dino"; + }; + /** + * HED Edge Detection + * @description Geneartes an edge map using the HED (softedge) model. + */ + HEDEdgeDetectionInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to process + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Scribble + * @description Whether or not to use scribble mode + * @default false + */ + scribble?: boolean; + /** + * type + * @default hed_edge_detection + * @constant + */ + type: "hed_edge_detection"; + }; + /** + * HFModelSource + * @description A HuggingFace repo_id with optional variant, sub-folder(s) and access token. + * Note that the variant option, if not provided to the constructor, will default to fp16, which is + * what people (almost) always want. + * + * The subfolder can be a single path or multiple paths joined by '+' (e.g., "text_encoder+tokenizer"). + * When multiple subfolders are specified, all of them will be downloaded and combined into the model directory. + */ + HFModelSource: { + /** Repo Id */ + repo_id: string; + /** @default fp16 */ + variant?: components["schemas"]["ModelRepoVariant"] | null; + /** Subfolder */ + subfolder?: string | null; + /** Access Token */ + access_token?: string | null; + /** + * @description discriminator enum property added by openapi-typescript + * @enum {string} + */ + type: "hf"; + }; + /** + * HFTokenStatus + * @enum {string} + */ + HFTokenStatus: "valid" | "invalid" | "unknown"; + /** HTTPValidationError */ + HTTPValidationError: { + /** Detail */ + detail?: components["schemas"]["ValidationError"][]; + }; + /** + * Halftone + * @description Halftones an image + */ + HalftoneInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to halftone + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Spacing + * @description Halftone dot spacing + * @default 8 + */ + spacing?: number; + /** + * Angle + * @description Halftone angle + * @default 45 + */ + angle?: number; + /** + * Oversampling + * @description Oversampling factor + * @default 1 + */ + oversampling?: number; + /** + * type + * @default halftone + * @constant + */ + type: "halftone"; + }; + /** + * HalvedPromptOutput + * @description Output for a halved prompt. + */ + HalvedPromptOutput: { + /** + * Prompt + * @description The entire output prompt + * @default + */ + prompt: string; + /** + * Part A + * @description The first part of the output prompt + * @default + */ + part_a: string; + /** + * Part B + * @description The second part of the output prompt + * @default + */ + part_b: string; + /** + * Resolutions + * @description JSON dict of [tagname,id] resolutions + * @default + */ + resolutions: string; + /** + * Seed Vector + * @description JSON string of the seed vector used for generation + * @default + */ + seed_vector: string; + /** + * type + * @default halved_prompt_output + * @constant + */ + type: "halved_prompt_output"; + }; + /** + * Heuristic Resize + * @description Resize an image using a heuristic method. Preserves edge maps. + */ + HeuristicResizeInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to resize + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Width + * @description The width to resize to (px) + * @default 512 + */ + width?: number; + /** + * Height + * @description The height to resize to (px) + * @default 512 + */ + height?: number; + /** + * type + * @default heuristic_resize + * @constant + */ + type: "heuristic_resize"; + }; + /** + * HuggingFaceMetadata + * @description Extended metadata fields provided by HuggingFace. + */ + HuggingFaceMetadata: { + /** + * Name + * @description model's name + */ + name: string; + /** + * Files + * @description model files and their sizes + */ + files?: components["schemas"]["RemoteModelFile"][]; + /** + * @description discriminator enum property added by openapi-typescript + * @enum {string} + */ + type: "huggingface"; + /** + * Id + * @description The HF model id + */ + id: string; + /** + * Api Response + * @description Response from the HF API as stringified JSON + */ + api_response?: string | null; + /** + * Is Diffusers + * @description Whether the metadata is for a Diffusers format model + * @default false + */ + is_diffusers?: boolean; + /** + * Ckpt Urls + * @description URLs for all checkpoint format models in the metadata + */ + ckpt_urls?: string[] | null; + }; + /** HuggingFaceModels */ + HuggingFaceModels: { + /** + * Urls + * @description URLs for all checkpoint format models in the metadata + */ + urls: string[] | null; + /** + * Is Diffusers + * @description Whether the metadata is for a Diffusers format model + */ + is_diffusers: boolean; + }; + /** IPAdapterField */ + IPAdapterField: { + /** + * Image + * @description The IP-Adapter image prompt(s). + */ + image: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][]; + /** @description The IP-Adapter model to use. */ + ip_adapter_model: components["schemas"]["ModelIdentifierField"]; + /** @description The name of the CLIP image encoder model. */ + image_encoder_model: components["schemas"]["ModelIdentifierField"]; + /** + * Weight + * @description The weight given to the IP-Adapter. + * @default 1 + */ + weight?: number | number[]; + /** + * Target Blocks + * @description The IP Adapter blocks to apply + * @default [] + */ + target_blocks?: string[]; + /** + * Method + * @description Weight apply method + * @default full + */ + method?: string; + /** + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) + * @default 0 + */ + begin_step_percent?: number; + /** + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; + /** + * @description The bool mask associated with this IP-Adapter. Excluded regions should be set to False, included regions should be set to True. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + }; + /** + * IP-Adapter - SD1.5, SDXL + * @description Collects IP-Adapter info to pass to other nodes. + */ + IPAdapterInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Image + * @description The IP-Adapter image prompt(s). + * @default null + */ + image?: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][] | null; + /** + * IP-Adapter Model + * @description The IP-Adapter model. + * @default null + */ + ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Clip Vision Model + * @description CLIP Vision model to use. Overrides model settings. Mandatory for checkpoint models. + * @default ViT-H + * @enum {string} + */ + clip_vision_model?: "ViT-H" | "ViT-G" | "ViT-L"; + /** + * Weight + * @description The weight given to the IP-Adapter + * @default 1 + */ + weight?: number | number[]; + /** + * Method + * @description The method to apply the IP-Adapter + * @default full + * @enum {string} + */ + method?: "full" | "style" | "composition" | "style_strong" | "style_precise"; + /** + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) + * @default 0 + */ + begin_step_percent?: number; + /** + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; + /** + * @description A mask defining the region that this IP-Adapter applies to. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + }; + /** + * IP-Adapter-Linked + * @description Collects IP-Adapter info to pass to other nodes. + */ + IPAdapterLinkedInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Image + * @description The IP-Adapter image prompt(s). + * @default null + */ + image?: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][] | null; + /** + * IP-Adapter Model + * @description The IP-Adapter model. + * @default null + */ + ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Clip Vision Model + * @description CLIP Vision model to use. Overrides model settings. Mandatory for checkpoint models. + * @default ViT-H + * @enum {string} + */ + clip_vision_model?: "ViT-H" | "ViT-G" | "ViT-L"; + /** + * Weight + * @description The weight given to the IP-Adapter + * @default 1 + */ + weight?: number | number[]; + /** + * Method + * @description The method to apply the IP-Adapter + * @default full + * @enum {string} + */ + method?: "full" | "style" | "composition" | "style_strong" | "style_precise"; + /** + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) + * @default 0 + */ + begin_step_percent?: number; + /** + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; + /** + * @description A mask defining the region that this IP-Adapter applies to. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * type + * @default ip_adapter_linked + * @constant + */ + type: "ip_adapter_linked"; + /** + * IP-Adapter List + * @description IP-Adapter to apply + * @default null + */ + ip_adapter_list?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + }; + /** IPAdapterListOutput */ + IPAdapterListOutput: { + /** + * IP-Adapter List + * @description IP-Adapter to apply + */ + ip_adapter_list: components["schemas"]["IPAdapterField"][]; + /** + * type + * @default ip_adapter_list_output + * @constant + */ + type: "ip_adapter_list_output"; + }; + /** + * IPAdapterMetadataField + * @description IP Adapter Field, minus the CLIP Vision Encoder model + */ + IPAdapterMetadataField: { + /** @description The IP-Adapter image prompt. */ + image: components["schemas"]["ImageField"]; + /** @description The IP-Adapter model. */ + ip_adapter_model: components["schemas"]["ModelIdentifierField"]; + /** + * Clip Vision Model + * @description The CLIP Vision model + * @enum {string} + */ + clip_vision_model: "ViT-L" | "ViT-H" | "ViT-G"; + /** + * Method + * @description Method to apply IP Weights with + * @enum {string} + */ + method: "full" | "style" | "composition" | "style_strong" | "style_precise"; + /** + * Weight + * @description The weight given to the IP-Adapter + */ + weight: number | number[]; + /** + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) + */ + begin_step_percent: number; + /** + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) + */ + end_step_percent: number; + }; + /** IPAdapterOutput */ + IPAdapterOutput: { + /** + * IP-Adapter + * @description IP-Adapter to apply + */ + ip_adapter: components["schemas"]["IPAdapterField"]; + /** + * type + * @default ip_adapter_output + * @constant + */ + type: "ip_adapter_output"; + }; + /** + * IPAdapterRecallParameter + * @description IP Adapter configuration for recall + */ + IPAdapterRecallParameter: { + /** + * Model Name + * @description The name of the IP Adapter model + */ + model_name: string; + /** + * Image Name + * @description The filename of the reference image in outputs/images + */ + image_name?: string | null; + /** + * Weight + * @description The weight for the IP Adapter + * @default 1 + */ + weight?: number; + /** + * Begin Step Percent + * @description When the IP Adapter is first applied (% of total steps) + */ + begin_step_percent?: number | null; + /** + * End Step Percent + * @description When the IP Adapter is last applied (% of total steps) + */ + end_step_percent?: number | null; + /** + * Method + * @description The IP Adapter method + */ + method?: ("full" | "style" | "composition") | null; + /** + * Image Influence + * @description FLUX Redux image influence (if model is flux_redux) + */ + image_influence?: ("lowest" | "low" | "medium" | "high" | "highest") | null; + }; + /** IPAdapter_Checkpoint_FLUX_Config */ + IPAdapter_Checkpoint_FLUX_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; + }; + /** IPAdapter_Checkpoint_SD1_Config */ + IPAdapter_Checkpoint_SD1_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + /** + * Base + * @default sd-1 + * @constant + */ + base: "sd-1"; + }; + /** IPAdapter_Checkpoint_SD2_Config */ + IPAdapter_Checkpoint_SD2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; + }; + /** IPAdapter_Checkpoint_SDXL_Config */ + IPAdapter_Checkpoint_SDXL_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; + }; + /** IPAdapter_InvokeAI_SD1_Config */ + IPAdapter_InvokeAI_SD1_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + /** + * Format + * @default invokeai + * @constant + */ + format: "invokeai"; + /** Image Encoder Model Id */ + image_encoder_model_id: string; + /** + * Base + * @default sd-1 + * @constant + */ + base: "sd-1"; + }; + /** IPAdapter_InvokeAI_SD2_Config */ + IPAdapter_InvokeAI_SD2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + /** + * Format + * @default invokeai + * @constant + */ + format: "invokeai"; + /** Image Encoder Model Id */ + image_encoder_model_id: string; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; + }; + /** IPAdapter_InvokeAI_SDXL_Config */ + IPAdapter_InvokeAI_SDXL_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + /** + * Format + * @default invokeai + * @constant + */ + format: "invokeai"; + /** Image Encoder Model Id */ + image_encoder_model_id: string; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; + }; + /** + * Ideal Size - SD1.5, SDXL + * @description Calculates the ideal size for generation to avoid duplication + */ + IdealSizeInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Width + * @description Final image width + * @default 1024 + */ + width?: number; + /** + * Height + * @description Final image height + * @default 576 + */ + height?: number; + /** + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet?: components["schemas"]["UNetField"] | null; + /** + * Multiplier + * @description Amount to multiply the model's dimensions by when calculating the ideal size (may result in initial generation artifacts if too large) + * @default 1 + */ + multiplier?: number; + /** + * type + * @default ideal_size + * @constant + */ + type: "ideal_size"; + }; + /** + * IdealSizeOutput + * @description Base class for invocations that output an image + */ + IdealSizeOutput: { + /** + * Width + * @description The ideal width of the image (in pixels) + */ + width: number; + /** + * Height + * @description The ideal height of the image (in pixels) + */ + height: number; + /** + * type + * @default ideal_size_output + * @constant + */ + type: "ideal_size_output"; + }; + /** + * Image Batch + * @description Create a batched generation, where the workflow is executed once for each image in the batch. + */ + ImageBatchInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Batch Group + * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. + * @default None + * @enum {string} + */ + batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; + /** + * Images + * @description The images to batch over + * @default null + */ + images?: components["schemas"]["ImageField"][] | null; + /** + * type + * @default image_batch + * @constant + */ + type: "image_batch"; + }; + /** + * Blur Image + * @description Blurs an image + */ + ImageBlurInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to blur + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Radius + * @description The blur radius + * @default 8 + */ + radius?: number; + /** + * Blur Type + * @description The type of blur + * @default gaussian + * @enum {string} + */ + blur_type?: "gaussian" | "box"; + /** + * type + * @default img_blur + * @constant + */ + type: "img_blur"; + }; + /** + * ImageCategory + * @description The category of an image. + * + * - GENERAL: The image is an output, init image, or otherwise an image without a specialized purpose. + * - MASK: The image is a mask image. + * - CONTROL: The image is a ControlNet control image. + * - USER: The image is a user-provide image. + * - OTHER: The image is some other type of image with a specialized purpose. To be used by external nodes. + * @enum {string} + */ + ImageCategory: "general" | "mask" | "control" | "user" | "other"; + /** + * Extract Image Channel + * @description Gets a channel from an image. + */ + ImageChannelInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to get the channel from + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Channel + * @description The channel to get + * @default A + * @enum {string} + */ + channel?: "A" | "R" | "G" | "B"; + /** + * type + * @default img_chan + * @constant + */ + type: "img_chan"; + }; + /** + * Multiply Image Channel + * @description Scale a specific color channel of an image. + */ + ImageChannelMultiplyInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to adjust + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Channel + * @description Which channel to adjust + * @default null + */ + channel?: ("Red (RGBA)" | "Green (RGBA)" | "Blue (RGBA)" | "Alpha (RGBA)" | "Cyan (CMYK)" | "Magenta (CMYK)" | "Yellow (CMYK)" | "Black (CMYK)" | "Hue (HSV)" | "Saturation (HSV)" | "Value (HSV)" | "Luminosity (LAB)" | "A (LAB)" | "B (LAB)" | "Y (YCbCr)" | "Cb (YCbCr)" | "Cr (YCbCr)") | null; + /** + * Scale + * @description The amount to scale the channel by. + * @default 1 + */ + scale?: number; + /** + * Invert Channel + * @description Invert the channel after scaling + * @default false + */ + invert_channel?: boolean; + /** + * type + * @default img_channel_multiply + * @constant + */ + type: "img_channel_multiply"; + }; + /** + * Offset Image Channel + * @description Add or subtract a value from a specific color channel of an image. + */ + ImageChannelOffsetInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to adjust + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Channel + * @description Which channel to adjust + * @default null + */ + channel?: ("Red (RGBA)" | "Green (RGBA)" | "Blue (RGBA)" | "Alpha (RGBA)" | "Cyan (CMYK)" | "Magenta (CMYK)" | "Yellow (CMYK)" | "Black (CMYK)" | "Hue (HSV)" | "Saturation (HSV)" | "Value (HSV)" | "Luminosity (LAB)" | "A (LAB)" | "B (LAB)" | "Y (YCbCr)" | "Cb (YCbCr)" | "Cr (YCbCr)") | null; + /** + * Offset + * @description The amount to adjust the channel by + * @default 0 + */ + offset?: number; + /** + * type + * @default img_channel_offset + * @constant + */ + type: "img_channel_offset"; + }; + /** + * Image Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option + */ + ImageCollectionIndexInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * Random + * @description Random Index? + * @default true + */ + random?: boolean; + /** + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 + */ + index?: number; + /** + * Collection + * @description image collection + * @default null + */ + collection?: components["schemas"]["ImageField"][] | null; + /** + * type + * @default image_collection_index + * @constant + */ + type: "image_collection_index"; + }; + /** + * Image Collection Primitive + * @description A collection of image primitive values + */ + ImageCollectionInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Collection + * @description The collection of image values + * @default null + */ + collection?: components["schemas"]["ImageField"][] | null; + /** + * type + * @default image_collection + * @constant + */ + type: "image_collection"; + }; + /** + * Image Collection Primitive linked + * @description A collection of image primitive values + */ + ImageCollectionLinkedInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Collection + * @description The collection of image values + * @default null + */ + collection?: components["schemas"]["ImageField"][] | null; + /** + * type + * @default image_collection_linked + * @constant + */ + type: "image_collection_linked"; + /** + * @description The image to load + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + }; + /** + * ImageCollectionOutput + * @description Base class for nodes that output a collection of images + */ + ImageCollectionOutput: { + /** + * Collection + * @description The output images + */ + collection: components["schemas"]["ImageField"][]; + /** + * type + * @default image_collection_output + * @constant + */ + type: "image_collection_output"; + }; + /** + * Image Collection Toggle + * @description Allows boolean selection between two separate Image collection inputs + */ + ImageCollectionToggleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * Col1 + * @description First Image Collection Input + * @default null + */ + col1?: components["schemas"]["ImageField"][] | null; + /** + * Col2 + * @description Second Image Collection Input + * @default null + */ + col2?: components["schemas"]["ImageField"][] | null; + /** + * type + * @default image_collection_toggle + * @constant + */ + type: "image_collection_toggle"; + }; + /** + * Convert Image Mode + * @description Converts an image to a different mode. + */ + ImageConvertInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to convert + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Mode + * @description The mode to convert to + * @default L + * @enum {string} + */ + mode?: "L" | "RGB" | "RGBA" | "CMYK" | "YCbCr" | "LAB" | "HSV" | "I" | "F"; + /** + * type + * @default img_conv + * @constant + */ + type: "img_conv"; + }; + /** + * Crop Image + * @description Crops an image to a specified box. The box can be outside of the image. + */ + ImageCropInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to crop + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * X + * @description The left x coordinate of the crop rectangle + * @default 0 + */ + x?: number; + /** + * Y + * @description The top y coordinate of the crop rectangle + * @default 0 + */ + y?: number; + /** + * Width + * @description The width of the crop rectangle + * @default 512 + */ + width?: number; + /** + * Height + * @description The height of the crop rectangle + * @default 512 + */ + height?: number; + /** + * type + * @default img_crop + * @constant + */ + type: "img_crop"; + }; + /** + * ImageDTO + * @description Deserialized image record, enriched for the frontend. + */ + ImageDTO: { + /** + * Image Name + * @description The unique name of the image. + */ + image_name: string; + /** + * Image Url + * @description The URL of the image. + */ + image_url: string; + /** + * Thumbnail Url + * @description The URL of the image's thumbnail. + */ + thumbnail_url: string; + /** @description The type of the image. */ + image_origin: components["schemas"]["ResourceOrigin"]; + /** @description The category of the image. */ + image_category: components["schemas"]["ImageCategory"]; + /** + * Width + * @description The width of the image in px. + */ + width: number; + /** + * Height + * @description The height of the image in px. + */ + height: number; + /** + * Created At + * @description The created timestamp of the image. + */ + created_at: string; + /** + * Updated At + * @description The updated timestamp of the image. + */ + updated_at: string; + /** + * Deleted At + * @description The deleted timestamp of the image. + */ + deleted_at?: string | null; + /** + * Is Intermediate + * @description Whether this is an intermediate image. + */ + is_intermediate: boolean; + /** + * Session Id + * @description The session ID that generated this image, if it is a generated image. + */ + session_id?: string | null; + /** + * Node Id + * @description The node ID that generated this image, if it is a generated image. + */ + node_id?: string | null; + /** + * Starred + * @description Whether this image is starred. + */ + starred: boolean; + /** + * Has Workflow + * @description Whether this image has a workflow. + */ + has_workflow: boolean; + /** + * Board Id + * @description The id of the board the image belongs to, if one exists. + */ + board_id?: string | null; + }; + /** + * ImageField + * @description An image primitive field + */ + ImageField: { + /** + * Image Name + * @description The name of the image + */ + image_name: string; + }; + /** + * Image Generator + * @description Generated a collection of images for use in a batched generation + */ + ImageGenerator: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Generator Type + * @description The image generator. + */ + generator: components["schemas"]["ImageGeneratorField"]; + /** + * type + * @default image_generator + * @constant + */ + type: "image_generator"; + }; + /** ImageGeneratorField */ + ImageGeneratorField: Record; + /** + * ImageGeneratorOutput + * @description Base class for nodes that output a collection of boards + */ + ImageGeneratorOutput: { + /** + * Images + * @description The generated images + */ + images: components["schemas"]["ImageField"][]; + /** + * type + * @default image_generator_output + * @constant + */ + type: "image_generator_output"; + }; + /** + * Adjust Image Hue + * @description Adjusts the Hue of an image. + */ + ImageHueAdjustmentInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to adjust + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Hue + * @description The degrees by which to rotate the hue, 0-360 + * @default 0 + */ + hue?: number; + /** + * type + * @default img_hue_adjust + * @constant + */ + type: "img_hue_adjust"; + }; + /** + * Image Index Collect + * @description ImageIndexCollect takes Image and Index then outputs it as an (index,image_name)array converted to json + */ + ImageIndexCollectInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Index + * @description The index + * @default 0 + */ + index?: number; + /** + * @description The image associated with the index + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default image_index_collect + * @constant + */ + type: "image_index_collect"; + }; + /** + * ImageIndexCollectOutput + * @description XImageCollectOutput string containing an array of xItem, Image_name converted to json + */ + ImageIndexCollectOutput: { + /** + * Image Index Collection + * @description The Image Index Collection + */ + image_index_collection: string; + /** + * type + * @default image_index_collect_output + * @constant + */ + type: "image_index_collect_output"; + }; + /** + * Inverse Lerp Image + * @description Inverse linear interpolation of all pixels of an image + */ + ImageInverseLerpInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to lerp + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Min + * @description The minimum input value + * @default 0 + */ + min?: number; + /** + * Max + * @description The maximum input value + * @default 255 + */ + max?: number; + /** + * type + * @default img_ilerp + * @constant + */ + type: "img_ilerp"; + }; + /** + * Image Primitive + * @description An image primitive value + */ + ImageInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to load + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default image + * @constant + */ + type: "image"; + }; + /** + * Lerp Image + * @description Linear interpolation of all pixels of an image + */ + ImageLerpInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to lerp + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Min + * @description The minimum output value + * @default 0 + */ + min?: number; + /** + * Max + * @description The maximum output value + * @default 255 + */ + max?: number; + /** + * type + * @default img_lerp + * @constant + */ + type: "img_lerp"; + }; + /** + * Image Mask to Tensor + * @description Convert a mask image to a tensor. Converts the image to grayscale and uses thresholding at the specified value. + */ + ImageMaskToTensorInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The mask image to convert. + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Cutoff + * @description Cutoff (<) + * @default 128 + */ + cutoff?: number; + /** + * Invert + * @description Whether to invert the mask. + * @default false + */ + invert?: boolean; + /** + * type + * @default image_mask_to_tensor + * @constant + */ + type: "image_mask_to_tensor"; + }; + /** + * Multiply Images + * @description Multiplies two images together using `PIL.ImageChops.multiply()`. + */ + ImageMultiplyInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The first image to multiply + * @default null + */ + image1?: components["schemas"]["ImageField"] | null; + /** + * @description The second image to multiply + * @default null + */ + image2?: components["schemas"]["ImageField"] | null; + /** + * type + * @default img_mul + * @constant + */ + type: "img_mul"; + }; + /** + * Blur NSFW Image + * @description Add blur to NSFW-flagged images + */ + ImageNSFWBlurInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to check + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default img_nsfw + * @constant + */ + type: "img_nsfw"; + }; + /** + * ImageNamesResult + * @description Response containing ordered image names with metadata for optimistic updates. + */ + ImageNamesResult: { + /** + * Image Names + * @description Ordered list of image names + */ + image_names: string[]; + /** + * Starred Count + * @description Number of starred images (when starred_first=True) + */ + starred_count: number; + /** + * Total Count + * @description Total number of images matching the query + */ + total_count: number; + }; + /** + * Add Image Noise + * @description Add noise to an image + */ + ImageNoiseInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to add noise to + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description Optional mask determining where to apply noise (black=noise, white=no noise) + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * Seed + * @description Seed for random number generation + * @default 0 + */ + seed?: number; + /** + * Noise Type + * @description The type of noise to add + * @default gaussian + * @enum {string} + */ + noise_type?: "gaussian" | "salt_and_pepper"; + /** + * Amount + * @description The amount of noise to add + * @default 0.1 + */ + amount?: number; + /** + * Noise Color + * @description Whether to add colored noise + * @default true + */ + noise_color?: boolean; + /** + * Size + * @description The size of the noise points + * @default 1 + */ + size?: number; + /** + * type + * @default img_noise + * @constant + */ + type: "img_noise"; + }; + /** + * Offset Image + * @description Offsets an image by a given percentage (or pixel amount). + * + * This works like Offset Latents, but in image space, with the additional capability of taking exact pixel offsets instead of just percentages (toggled with a switch/boolean input). + */ + ImageOffsetInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * As Pixels + * @description Interpret offsets as pixels rather than percentages + * @default false + */ + as_pixels?: boolean; + /** + * @description Image to be offset + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * X Offset + * @description x-offset for the subject + * @default 0.5 + */ + x_offset?: number; + /** + * Y Offset + * @description y-offset for the subject + * @default 0.5 + */ + y_offset?: number; + /** + * type + * @default offset_image + * @constant + */ + type: "offset_image"; + }; + /** + * ImageOutput + * @description Base class for nodes that output a single image + */ + ImageOutput: { + /** @description The output image */ + image: components["schemas"]["ImageField"]; + /** + * Width + * @description The width of the image in pixels + */ + width: number; + /** + * Height + * @description The height of the image in pixels + */ + height: number; + /** + * type + * @default image_output + * @constant + */ + type: "image_output"; + }; + /** ImagePanelCoordinateOutput */ + ImagePanelCoordinateOutput: { + /** + * X Left + * @description The left x-coordinate of the panel. + */ + x_left: number; + /** + * Y Top + * @description The top y-coordinate of the panel. + */ + y_top: number; + /** + * Width + * @description The width of the panel. + */ + width: number; + /** + * Height + * @description The height of the panel. + */ + height: number; + /** + * type + * @default image_panel_coordinate_output + * @constant + */ + type: "image_panel_coordinate_output"; + }; + /** + * Image Panel Layout + * @description Get the coordinates of a single panel in a grid. (If the full image shape cannot be divided evenly into panels, + * then the grid may not cover the entire image.) + */ + ImagePanelLayoutInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Width + * @description The width of the entire grid. + * @default null + */ + width?: number | null; + /** + * Height + * @description The height of the entire grid. + * @default null + */ + height?: number | null; + /** + * Num Cols + * @description The number of columns in the grid. + * @default 1 + */ + num_cols?: number; + /** + * Num Rows + * @description The number of rows in the grid. + * @default 1 + */ + num_rows?: number; + /** + * Panel Col Idx + * @description The column index of the panel to be processed. + * @default 0 + */ + panel_col_idx?: number; + /** + * Panel Row Idx + * @description The row index of the panel to be processed. + * @default 0 + */ + panel_row_idx?: number; + /** + * type + * @default image_panel_layout + * @constant + */ + type: "image_panel_layout"; + }; + /** + * Paste Image + * @description Pastes an image into another image. + */ + ImagePasteInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The base image + * @default null + */ + base_image?: components["schemas"]["ImageField"] | null; + /** + * @description The image to paste + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description The mask to use when pasting + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * X + * @description The left x coordinate at which to paste the image + * @default 0 + */ + x?: number; + /** + * Y + * @description The top y coordinate at which to paste the image + * @default 0 + */ + y?: number; + /** + * Crop + * @description Crop to base image dimensions + * @default false + */ + crop?: boolean; + /** + * type + * @default img_paste + * @constant + */ + type: "img_paste"; + }; + /** + * ImageRecordChanges + * @description A set of changes to apply to an image record. + * + * Only limited changes are valid: + * - `image_category`: change the category of an image + * - `session_id`: change the session associated with an image + * - `is_intermediate`: change the image's `is_intermediate` flag + * - `starred`: change whether the image is starred + */ + ImageRecordChanges: { + /** @description The image's new category. */ + image_category?: components["schemas"]["ImageCategory"] | null; + /** + * Session Id + * @description The image's new session ID. + */ + session_id?: string | null; + /** + * Is Intermediate + * @description The image's new `is_intermediate` flag. + */ + is_intermediate?: boolean | null; + /** + * Starred + * @description The image's new `starred` state + */ + starred?: boolean | null; + } & { + [key: string]: unknown; + }; + /** + * Resize Image + * @description Resizes an image to specific dimensions + */ + ImageResizeInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to resize + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Width + * @description The width to resize to (px) + * @default 512 + */ + width?: number; + /** + * Height + * @description The height to resize to (px) + * @default 512 + */ + height?: number; + /** + * Resample Mode + * @description The resampling mode + * @default bicubic + * @enum {string} + */ + resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; + /** + * type + * @default img_resize + * @constant + */ + type: "img_resize"; + }; + /** + * Rotate/Flip Image + * @description Rotates an image by a given angle (in degrees clockwise). + * + * Rotate an image in degrees about its center, clockwise (positive entries) or counterclockwise (negative entries). Optionally expand the image boundary to fit the rotated image, or flip it horizontally or vertically. + */ + ImageRotateInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Image to be rotated clockwise + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Degrees + * @description Angle (in degrees clockwise) by which to rotate + * @default 90 + */ + degrees?: number; + /** + * Expand To Fit + * @description If true, extends the image boundary to fit the rotated content + * @default true + */ + expand_to_fit?: boolean; + /** + * Flip Horizontal + * @description If true, flips the image horizontally + * @default false + */ + flip_horizontal?: boolean; + /** + * Flip Vertical + * @description If true, flips the image vertically + * @default false + */ + flip_vertical?: boolean; + /** + * type + * @default rotate_image + * @constant + */ + type: "rotate_image"; + }; + /** + * Image Quantize (Kohonen map) + * @description Use a Kohonen self-organizing map to quantize the pixel values of an image. + * + * It's possible to pass in an existing map, which will be used instead of training a new one. It's also possible to pass in a "swap map", which will be used in place of the standard map's assigned pixel values in quantizing the target image - these values can be correlated either one by one by a linear assignment minimizing the distances\* between each of them, or by swapping their coordinates on the maps themselves, which can be oriented first such that their corner distances\* are minimized achieving a closest-fit while attempting to preserve mappings of adjacent colors. + * + * \*Here, "distances" refers to the euclidean distances between (L, a, b) tuples in Oklab color space. + */ + ImageSOMInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to quantize + * @default null + */ + image_in?: components["schemas"]["ImageField"] | null; + /** + * @description Use an existing SOM instead of training one (skips all training) + * @default null + */ + map_in?: components["schemas"]["ImageField"] | null; + /** + * @description Take another map and swap in its colors after obtaining best-match indices but prior to mapping + * @default null + */ + swap_map?: components["schemas"]["ImageField"] | null; + /** + * Swap Mode + * @description How to employ the swap map - directly, reoriented or rearranged + * @default Minimize distances + * @enum {string} + */ + swap_mode?: "Direct" | "Reorient corners" | "Minimize distances"; + /** + * Map Width + * @description Width (in cells) of the self-organizing map to train + * @default 16 + */ + map_width?: number; + /** + * Map Height + * @description Height (in cells) of the self-organizing map to train + * @default 16 + */ + map_height?: number; + /** + * Steps + * @description Training step count for the self-organizing map + * @default 64 + */ + steps?: number; + /** + * Training Scale + * @description Nearest-neighbor scale image size prior to sampling - size close to sample size is recommended + * @default 0.25 + */ + training_scale?: number; + /** + * Sample Width + * @description Width of assorted pixel sample per step - for performance, keep this number low + * @default 64 + */ + sample_width?: number; + /** + * Sample Height + * @description Height of assorted pixel sample per step - for performance, keep this number low + * @default 64 + */ + sample_height?: number; + /** + * type + * @default image_som + * @constant + */ + type: "image_som"; + }; + /** + * ImageSOMOutput + * @description Outputs an image and the SOM used to quantize that image + */ + ImageSOMOutput: { + /** @description Quantized image */ + image_out: components["schemas"]["ImageField"]; + /** @description The pixels of the self-organizing map */ + map_out: components["schemas"]["ImageField"]; + /** + * Image Width + * @description Width of the quantized image + */ + image_width: number; + /** + * Image Height + * @description Height of the quantized image + */ + image_height: number; + /** + * Map Width + * @description Width of the SOM image + */ + map_width: number; + /** + * Map Height + * @description Height of the SOM image + */ + map_height: number; + /** + * type + * @default image_som_output + * @constant + */ + type: "image_som_output"; + }; + /** + * Scale Image + * @description Scales an image by a factor + */ + ImageScaleInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to scale + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Scale Factor + * @description The factor by which to scale the image + * @default 2 + */ + scale_factor?: number; + /** + * Resample Mode + * @description The resampling mode + * @default bicubic + * @enum {string} + */ + resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; + /** + * type + * @default img_scale + * @constant + */ + type: "img_scale"; + }; + /** + * Image Search to Mask (Clipseg) + * @description Uses the Clipseg model to generate an image mask from an image prompt. + * + * Input a base image and a prompt image to generate a mask representing areas of the base image matched by the prompt image contents. + */ + ImageSearchToMaskClipsegInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image from which to create a mask + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description Image prompt for which to search + * @default null + */ + search_image?: components["schemas"]["ImageField"] | null; + /** + * Invert Output + * @description Off: white on black / On: black on white + * @default true + */ + invert_output?: boolean; + /** + * Smoothing + * @description Radius of blur to apply before thresholding + * @default 4 + */ + smoothing?: number; + /** + * Subject Threshold + * @description Threshold above which is considered the subject + * @default 0.4 + */ + subject_threshold?: number; + /** + * Background Threshold + * @description Threshold below which is considered the background + * @default 0.4 + */ + background_threshold?: number; + /** + * Mask Expand Or Contract + * @description Pixels by which to grow (or shrink) mask after thresholding + * @default 0 + */ + mask_expand_or_contract?: number; + /** + * Mask Blur + * @description Radius of blur to apply after thresholding + * @default 0 + */ + mask_blur?: number; + /** + * type + * @default imgs2mask_clipseg + * @constant + */ + type: "imgs2mask_clipseg"; + }; + /** + * Image to ASCII Art AnyFont + * @description Convert an Image to Ascii Art Image using any font or size + * https://github.com/dernyn/256/tree/master this is a great font to use + */ + ImageToAAInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * @description Image to convert to ASCII art + * @default null + */ + input_image?: components["schemas"]["ImageField"] | null; + /** + * Font Url + * @description URL address of the font file to download + * @default https://github.com/dernyn/256/raw/master/Dernyn's-256(baseline).ttf + */ + font_url?: string | null; + /** + * Local Font Path + * @description Local font file path (overrides font_url) + * @default null + */ + local_font_path?: string | null; + /** + * Local Font + * @description Name of the local font file to use from the font_cache folder + * @default None + * @constant + */ + local_font?: "None"; + /** + * Font Size + * @description Font size for the ASCII art characters + * @default 6 + */ + font_size?: number; + /** + * Character Range + * @description The character range to use + * @default Ascii + * @enum {string} + */ + character_range?: "All" | "Low" | "High" | "Ascii" | "Numbers" | "Letters" | "Lowercase" | "Uppercase" | "Hex" | "Punctuation" | "Printable" | "AH" | "AM" | "AL" | "Blocks" | "Binary" | "Custom"; + /** + * Custom Characters + * @description Custom characters. Used if Custom is selected from character range + * @default ▁▂▃▄▅▆▇█ + */ + custom_characters?: string; + /** + * Comparison Type + * @description Choose the comparison type (Sum of Absolute Differences, Mean Squared Error, Structural Similarity Index, Normalized Average Luminance) + * @default NAL + * @enum {string} + */ + comparison_type?: "SAD" | "MSE" | "SSIM" | "NAL"; + /** + * Mono Comparison + * @description Convert input image to mono for comparison + * @default false + */ + mono_comparison?: boolean; + /** + * Color Mode + * @description Enable color mode (default: grayscale) + * @default false + */ + color_mode?: boolean; + /** + * type + * @default I2AA_AnyFont + * @constant + */ + type: "I2AA_AnyFont"; + }; + /** + * Image to ASCII Art Image + * @description Convert an Image to ASCII Art Image + */ + ImageToDetailedASCIIArtInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * @description Input image to convert to ASCII art + * @default null + */ + input_image?: components["schemas"]["ImageField"] | null; + /** + * Font Spacing + * @description Font spacing for the ASCII art characters + * @default 6 + */ + font_spacing?: number; + /** + * Ascii Set + * @description Choose the desired ASCII character set + * @default Medium Detail + * @enum {string} + */ + ascii_set?: "High Detail" | "Medium Detail" | "Low Detail" | "Numbers" | "Blocks" | "Binary"; + /** + * Color Mode + * @description Enable color mode (default: grayscale) + * @default false + */ + color_mode?: boolean; + /** + * Invert Colors + * @description Invert background color and ASCII character order + * @default true + */ + invert_colors?: boolean; + /** + * Output To File + * @description Output ASCII art to a text file + * @default false + */ + output_to_file?: boolean; + /** + * Gamma + * @description Gamma correction value for the output image + * @default 1 + */ + gamma?: number; + /** + * type + * @default Image_to_ASCII_Art_Image + * @constant + */ + type: "Image_to_ASCII_Art_Image"; + }; + /** + * Image to Image Name + * @description Invocation to extract the image name from an ImageField. + */ + ImageToImageNameInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The ImageField to extract the name from + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default image_to_image_name + * @constant + */ + type: "image_to_image_name"; + }; + /** + * ImageToImageNameOutput + * @description Output class for Image to Image Name Invocation + */ + ImageToImageNameOutput: { + /** + * Image Name + * @description The name of the image + */ + image_name: string; + /** + * type + * @default image_to_image_name_output + * @constant + */ + type: "image_to_image_name_output"; + }; + /** + * Image to Latents - SD1.5, SDXL + * @description Encodes an image into latents. + */ + ImageToLatentsInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to encode + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description VAE + * @default null + */ + vae?: components["schemas"]["VAEField"] | null; + /** + * Tiled + * @description Processing using overlapping tiles (reduce memory consumption) + * @default false + */ + tiled?: boolean; + /** + * Tile Size + * @description The tile size for VAE tiling in pixels (image space). If set to 0, the default tile size for the model will be used. Larger tile sizes generally produce better results at the cost of higher memory usage. + * @default 0 + */ + tile_size?: number; + /** + * Fp32 + * @description Whether or not to use full float32 precision + * @default false + */ + fp32?: boolean; + /** + * Color Compensation + * @description Apply VAE scaling compensation when encoding images (reduces color drift). + * @default None + * @enum {string} + */ + color_compensation?: "None" | "SDXL"; + /** + * type + * @default i2l + * @constant + */ + type: "i2l"; + }; + /** + * Image to Unicode Art + * @description Convert an Image to Unicode Art using Extended Characters + */ + ImageToUnicodeArtInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * @description Input image to convert to Unicode art + * @default null + */ + input_image?: components["schemas"]["ImageField"] | null; + /** + * Font Size + * @description Font size for the Unicode art characters + * @default 8 + */ + font_size?: number; + /** + * Gamma + * @description Gamma correction value for the output image + * @default 1 + */ + gamma?: number; + /** + * Unicode Set + * @description Use shaded Unicode characters for artwork + * @default Shaded + * @enum {string} + */ + unicode_set?: "Shaded" | "Extended Shading" | "Intermediate Detail" | "Checkerboard Patterns" | "Vertical Lines" | "Horizontal Lines" | "Diagonal Lines" | "Arrows" | "Circles" | "Blocks" | "Triangles" | "Math Symbols" | "Stars"; + /** + * Color Mode + * @description Enable color mode (default: grayscale) + * @default true + */ + color_mode?: boolean; + /** + * Invert Colors + * @description Invert background color and ASCII character order + * @default true + */ + invert_colors?: boolean; + /** + * type + * @default Image_to_Unicode_Art + * @constant + */ + type: "Image_to_Unicode_Art"; + }; + /** + * Image To XYImage Collection + * @description Cuts an image up into columns and rows and outputs XYImage Collection + */ + ImageToXYImageCollectionInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The input image + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Columns + * @description The number of columns + * @default 2 + */ + columns?: number; + /** + * Rows + * @description The number of rows + * @default 2 + */ + rows?: number; + /** + * type + * @default image_to_xy_image_collection + * @constant + */ + type: "image_to_xy_image_collection"; + }; + /** + * Image To XYImage Tiles + * @description Cuts an image up into overlapping tiles and outputs as an XYImage Collection (x,y is the final position of the tile) + */ + ImageToXYImageTilesInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Tiles + * @description The list of tiles + * @default [] + */ + tiles?: string[]; + /** + * type + * @default image_to_xy_image_tiles + * @constant + */ + type: "image_to_xy_image_tiles"; + }; + /** + * ImageToXYImageTilesOutput + * @description Image To XYImage Tiles Output + */ + ImageToXYImageTilesOutput: { + /** + * Xyimages + * @description The XYImage Collection + */ + xyImages: string[]; + /** + * type + * @default image_to_xy_image_output + * @constant + */ + type: "image_to_xy_image_output"; + }; + /** + * Image Toggle + * @description Allows boolean selection between two separate image inputs + */ + ImageToggleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * @description First Image Input + * @default null + */ + img1?: components["schemas"]["ImageField"] | null; + /** + * @description Second Image Input + * @default null + */ + img2?: components["schemas"]["ImageField"] | null; + /** + * type + * @default image_toggle + * @constant + */ + type: "image_toggle"; + }; + /** ImageUploadEntry */ + ImageUploadEntry: { + /** @description The image DTO */ + image_dto: components["schemas"]["ImageDTO"]; + /** + * Presigned Url + * @description The URL to get the presigned URL for the image upload + */ + presigned_url: string; + }; + /** + * ImageUrlsDTO + * @description The URLs for an image and its thumbnail. + */ + ImageUrlsDTO: { + /** + * Image Name + * @description The unique name of the image. + */ + image_name: string; + /** + * Image Url + * @description The URL of the image. + */ + image_url: string; + /** + * Thumbnail Url + * @description The URL of the image's thumbnail. + */ + thumbnail_url: string; + }; + /** + * Add Invisible Watermark + * @description Add an invisible watermark to an image + */ + ImageWatermarkInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to check + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Text + * @description Watermark text + * @default InvokeAI + */ + text?: string; + /** + * type + * @default img_watermark + * @constant + */ + type: "img_watermark"; + }; + /** ImagesDownloaded */ + ImagesDownloaded: { + /** + * Response + * @description The message to display to the user when images begin downloading + */ + response?: string | null; + /** + * Bulk Download Item Name + * @description The name of the bulk download item for which events will be emitted + */ + bulk_download_item_name?: string | null; + }; + /** + * Images Index To Video Output + * @description Load a collection of ImageIndex types (json of (idex,image_name)array) and create a video of them + */ + ImagesIndexToVideoInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Image Index Collection + * @description The Image Index Collection + * @default null + */ + image_index_collection?: string[] | null; + /** + * Video Out Path + * @description Path and filename of output mp4 + * @default + */ + video_out_path?: string; + /** + * Fps + * @description FPS + * @default 30 + */ + fps?: number; + /** + * Codec + * @description Video codec FourCC (e.g., mp4v, avc1, h264, x264, hev1, vp09, av01) + * @default x264 + */ + codec?: string; + /** + * type + * @default images_index_to_video + * @constant + */ + type: "images_index_to_video"; + }; + /** + * ImagesIndexToVideoOutput + * @description ImagesIndexToVideoOutput returns nothing + */ + ImagesIndexToVideoOutput: { + /** + * type + * @default ImagesIndexToVideoOutput + * @constant + */ + type: "ImagesIndexToVideoOutput"; + }; + /** + * Images To Grids + * @description Takes a collection of images and outputs a collection of generated grid images + */ + ImagesToGridsInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Images + * @description The image collection to turn into grids + * @default [] + */ + images?: components["schemas"]["ImageField"][]; + /** + * Columns + * @description The number of columns in each grid + * @default 1 + */ + columns?: number; + /** + * Rows + * @description The number of rows to have in each grid + * @default 1 + */ + rows?: number; + /** + * Space + * @description The space to be added between images + * @default 1 + */ + space?: number; + /** + * Scale Factor + * @description The factor by which to scale the images + * @default 1 + */ + scale_factor?: number; + /** + * Resample Mode + * @description The resampling mode + * @default bicubic + * @enum {string} + */ + resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; + /** + * @description The color to use as the background + * @default { + * "r": 0, + * "g": 0, + * "b": 0, + * "a": 255 + * } + */ + background_color?: components["schemas"]["ColorField"]; + /** + * type + * @default images_to_grids + * @constant + */ + type: "images_to_grids"; + }; + /** + * Solid Color Infill + * @description Infills transparent areas of an image with a solid color + */ + InfillColorInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to process + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description The color to use to infill + * @default { + * "r": 127, + * "g": 127, + * "b": 127, + * "a": 255 + * } + */ + color?: components["schemas"]["ColorField"]; + /** + * type + * @default infill_rgba + * @constant + */ + type: "infill_rgba"; + }; + /** + * PatchMatch Infill + * @description Infills transparent areas of an image using the PatchMatch algorithm + */ + InfillPatchMatchInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to process + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Downscale + * @description Run patchmatch on downscaled image to speedup infill + * @default 2 + */ + downscale?: number; + /** + * Resample Mode + * @description The resampling mode + * @default bicubic + * @enum {string} + */ + resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; + /** + * type + * @default infill_patchmatch + * @constant + */ + type: "infill_patchmatch"; + }; + /** + * Tile Infill + * @description Infills transparent areas of an image with tiles of the image + */ + InfillTileInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to process + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Tile Size + * @description The tile size (px) + * @default 32 + */ + tile_size?: number; + /** + * Seed + * @description The seed to use for tile generation (omit for random) + * @default 0 + */ + seed?: number; + /** + * type + * @default infill_tile + * @constant + */ + type: "infill_tile"; + }; + /** + * UNet Info Grabber + * @description Outputs different info from a UNet + */ + InfoGrabberUNetInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * UNet + * @default null + */ + unet?: components["schemas"]["UNetField"] | null; + /** + * Info Type + * @description The kind of info to retrieve + * @default null + */ + info_type?: ("name" | "path") | null; + /** + * type + * @default info_grabber_unet_invocation + * @constant + */ + type: "info_grabber_unet_invocation"; + }; + /** + * Input + * @description The type of input a field accepts. + * - `Input.Direct`: The field must have its value provided directly, when the invocation and field are instantiated. + * - `Input.Connection`: The field must have its value provided by a connection. + * - `Input.Any`: The field may have its value provided either directly or by a connection. + * @enum {string} + */ + Input: "connection" | "direct" | "any"; + /** + * InputFieldJSONSchemaExtra + * @description Extra attributes to be added to input fields and their OpenAPI schema. Used during graph execution, + * and by the workflow editor during schema parsing and UI rendering. + */ + InputFieldJSONSchemaExtra: { + input: components["schemas"]["Input"]; + field_kind: components["schemas"]["FieldKind"]; + /** + * Orig Required + * @default true + */ + orig_required: boolean; + /** + * Default + * @default null + */ + default: unknown | null; + /** + * Orig Default + * @default null + */ + orig_default: unknown | null; + /** + * Ui Hidden + * @default false + */ + ui_hidden: boolean; + /** @default null */ + ui_type: components["schemas"]["UIType"] | null; + /** @default null */ + ui_component: components["schemas"]["UIComponent"] | null; + /** + * Ui Order + * @default null + */ + ui_order: number | null; + /** + * Ui Choice Labels + * @default null + */ + ui_choice_labels: { + [key: string]: string; + } | null; + /** + * Ui Model Base + * @default null + */ + ui_model_base: components["schemas"]["BaseModelType"][] | null; + /** + * Ui Model Type + * @default null + */ + ui_model_type: components["schemas"]["ModelType"][] | null; + /** + * Ui Model Variant + * @default null + */ + ui_model_variant: (components["schemas"]["ClipVariantType"] | components["schemas"]["ModelVariantType"])[] | null; + /** + * Ui Model Format + * @default null + */ + ui_model_format: components["schemas"]["ModelFormat"][] | null; + }; + /** + * InstallStatus + * @description State of an install job running in the background. + * @enum {string} + */ + InstallStatus: "waiting" | "downloading" | "downloads_done" | "running" | "paused" | "completed" | "error" | "cancelled"; + /** + * Integer Collection Toggle + * @description Allows boolean selection between two separate integer collection inputs + */ + IntCollectionToggleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * Col1 + * @description First Int Collection Input + * @default null + */ + col1?: number[] | null; + /** + * Col2 + * @description Second Int Collection Input + * @default null + */ + col2?: number[] | null; + /** + * type + * @default int_collection_toggle + * @constant + */ + type: "int_collection_toggle"; + }; + /** + * Integer Toggle + * @description Allows boolean selection between two separate integer inputs + */ + IntToggleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * Int1 + * @description First Integer Input + * @default 0 + */ + int1?: number; + /** + * Int2 + * @description Second Integer Input + * @default 0 + */ + int2?: number; + /** + * type + * @default int_toggle + * @constant + */ + type: "int_toggle"; + }; + /** + * Integer Batch + * @description Create a batched generation, where the workflow is executed once for each integer in the batch. + */ + IntegerBatchInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Batch Group + * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. + * @default None + * @enum {string} + */ + batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; + /** + * Integers + * @description The integers to batch over + * @default null + */ + integers?: number[] | null; + /** + * type + * @default integer_batch + * @constant + */ + type: "integer_batch"; + }; + /** + * Integer Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option + */ + IntegerCollectionIndexInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * Random + * @description Random Index? + * @default true + */ + random?: boolean; + /** + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 + */ + index?: number; + /** + * Collection + * @description integer collection + * @default null + */ + collection?: number[] | null; + /** + * type + * @default integer_collection_index + * @constant + */ + type: "integer_collection_index"; + }; + /** + * Integer Collection Primitive + * @description A collection of integer primitive values + */ + IntegerCollectionInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Collection + * @description The collection of integer values + * @default [] + */ + collection?: number[]; + /** + * type + * @default integer_collection + * @constant + */ + type: "integer_collection"; + }; + /** + * Integer Collection Primitive Linked + * @description A collection of integer primitive values + */ + IntegerCollectionLinkedInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Collection + * @description The collection of integer values + * @default [] + */ + collection?: number[]; + /** + * type + * @default integer_collection_linked + * @constant + */ + type: "integer_collection_linked"; + /** + * Value + * @description The integer value + * @default 0 + */ + value?: number; + }; + /** + * IntegerCollectionOutput + * @description Base class for nodes that output a collection of integers + */ + IntegerCollectionOutput: { + /** + * Collection + * @description The int collection + */ + collection: number[]; + /** + * type + * @default integer_collection_output + * @constant + */ + type: "integer_collection_output"; + }; + /** + * Integer Generator + * @description Generated a range of integers for use in a batched generation + */ + IntegerGenerator: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Generator Type + * @description The integer generator. + */ + generator: components["schemas"]["IntegerGeneratorField"]; + /** + * type + * @default integer_generator + * @constant + */ + type: "integer_generator"; + }; + /** IntegerGeneratorField */ + IntegerGeneratorField: Record; + /** IntegerGeneratorOutput */ + IntegerGeneratorOutput: { + /** + * Integers + * @description The generated integers + */ + integers: number[]; + /** + * type + * @default integer_generator_output + * @constant + */ + type: "integer_generator_output"; + }; + /** + * Integer Primitive + * @description An integer primitive value + */ + IntegerInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Value + * @description The integer value + * @default 0 + */ + value?: number; + /** + * type + * @default integer + * @constant + */ + type: "integer"; + }; + /** + * Integer Math + * @description Performs integer math. + */ + IntegerMathInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Operation + * @description The operation to perform + * @default ADD + * @enum {string} + */ + operation?: "ADD" | "SUB" | "MUL" | "DIV" | "EXP" | "MOD" | "ABS" | "MIN" | "MAX"; + /** + * A + * @description The first number + * @default 1 + */ + a?: number; + /** + * B + * @description The second number + * @default 1 + */ + b?: number; + /** + * type + * @default integer_math + * @constant + */ + type: "integer_math"; + }; + /** + * IntegerOutput + * @description Base class for nodes that output a single integer + */ + IntegerOutput: { + /** + * Value + * @description The output integer + */ + value: number; + /** + * type + * @default integer_output + * @constant + */ + type: "integer_output"; + }; + /** + * Ints To Strings + * @description Converts an integer or collection of integers to a collection of strings + */ + IntsToStringsInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Ints + * @description int or collection of ints + * @default [] + */ + ints?: number | number[]; + /** + * type + * @default ints_to_strings + * @constant + */ + type: "ints_to_strings"; + }; + /** + * Invert Tensor Mask + * @description Inverts a tensor mask. + */ + InvertTensorMaskInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The tensor mask to convert. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * type + * @default invert_tensor_mask + * @constant + */ + type: "invert_tensor_mask"; + }; + /** InvocationCacheStatus */ + InvocationCacheStatus: { + /** + * Size + * @description The current size of the invocation cache + */ + size: number; + /** + * Hits + * @description The number of cache hits + */ + hits: number; + /** + * Misses + * @description The number of cache misses + */ + misses: number; + /** + * Enabled + * @description Whether the invocation cache is enabled + */ + enabled: boolean; + /** + * Max Size + * @description The maximum size of the invocation cache + */ + max_size: number; + }; + /** + * InvocationCompleteEvent + * @description Event model for invocation_complete + */ + InvocationCompleteEvent: { + /** + * Timestamp + * @description The timestamp of the event + */ + timestamp: number; + /** + * Queue Id + * @description The ID of the queue + */ + queue_id: string; + /** + * Item Id + * @description The ID of the queue item + */ + item_id: number; + /** + * Batch Id + * @description The ID of the queue batch + */ + batch_id: string; + /** + * Origin + * @description The origin of the queue item + * @default null + */ + origin: string | null; + /** + * Destination + * @description The destination of the queue item + * @default null + */ + destination: string | null; + /** + * User Id + * @description The ID of the user who created the queue item + * @default system + */ + user_id: string; + /** + * Session Id + * @description The ID of the session (aka graph execution state) + */ + session_id: string; + /** + * Invocation + * @description The ID of the invocation + */ + invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + /** + * Invocation Source Id + * @description The ID of the prepared invocation's source node + */ + invocation_source_id: string; + /** + * Result + * @description The result of the invocation + */ + result: components["schemas"]["BooleanCollectionOutput"] | components["schemas"]["BooleanOutput"] | components["schemas"]["BoundingBoxCollectionOutput"] | components["schemas"]["BoundingBoxOutput"] | components["schemas"]["CLIPOutput"] | components["schemas"]["CLIPSkipInvocationOutput"] | components["schemas"]["CMYKSeparationOutput"] | components["schemas"]["CMYKSplitOutput"] | components["schemas"]["CalculateImageTilesOutput"] | components["schemas"]["ClipsegMaskHierarchyOutput"] | components["schemas"]["CogView4ConditioningOutput"] | components["schemas"]["CogView4ModelLoaderOutput"] | components["schemas"]["CollectInvocationOutput"] | components["schemas"]["CollectionCountOutput"] | components["schemas"]["CollectionIndexOutput"] | components["schemas"]["CollectionJoinOutput"] | components["schemas"]["CollectionReverseOutput"] | components["schemas"]["CollectionSliceOutput"] | components["schemas"]["CollectionSortOutput"] | components["schemas"]["CollectionUniqueOutput"] | components["schemas"]["ColorCollectionOutput"] | components["schemas"]["ColorOutput"] | components["schemas"]["ConditioningCollectionOutput"] | components["schemas"]["ConditioningOutput"] | components["schemas"]["ControlListOutput"] | components["schemas"]["ControlOutput"] | components["schemas"]["DenoiseMaskOutput"] | components["schemas"]["EscapedOutput"] | components["schemas"]["EvolutionaryPromptListOutput"] | components["schemas"]["FaceMaskOutput"] | components["schemas"]["FaceOffOutput"] | components["schemas"]["FloatCollectionOutput"] | components["schemas"]["FloatGeneratorOutput"] | components["schemas"]["FloatOutput"] | components["schemas"]["Flux2KleinLoRALoaderOutput"] | components["schemas"]["Flux2KleinModelLoaderOutput"] | components["schemas"]["FluxConditioningBlendOutput"] | components["schemas"]["FluxConditioningCollectionOutput"] | components["schemas"]["FluxConditioningDeltaAndAugmentedOutput"] | components["schemas"]["FluxConditioningListOutput"] | components["schemas"]["FluxConditioningOutput"] | components["schemas"]["FluxConditioningStoreOutput"] | components["schemas"]["FluxControlLoRALoaderOutput"] | components["schemas"]["FluxControlNetCollectionOutput"] | components["schemas"]["FluxControlNetListOutput"] | components["schemas"]["FluxControlNetOutput"] | components["schemas"]["FluxFillOutput"] | components["schemas"]["FluxIdealSizeOutput"] | components["schemas"]["FluxKontextOutput"] | components["schemas"]["FluxLoRALoaderOutput"] | components["schemas"]["FluxModelLoaderOutput"] | components["schemas"]["FluxReduxCollectionOutput"] | components["schemas"]["FluxReduxListOutput"] | components["schemas"]["FluxReduxOutput"] | components["schemas"]["FluxTextEncoderListOutput"] | components["schemas"]["FluxWeightedPromptOutput"] | components["schemas"]["GradientMaskOutput"] | components["schemas"]["HalvedPromptOutput"] | components["schemas"]["IPAdapterListOutput"] | components["schemas"]["IPAdapterOutput"] | components["schemas"]["IdealSizeOutput"] | components["schemas"]["ImageCollectionOutput"] | components["schemas"]["ImageGeneratorOutput"] | components["schemas"]["ImageIndexCollectOutput"] | components["schemas"]["ImageOutput"] | components["schemas"]["ImagePanelCoordinateOutput"] | components["schemas"]["ImageSOMOutput"] | components["schemas"]["ImageToImageNameOutput"] | components["schemas"]["ImageToXYImageTilesOutput"] | components["schemas"]["ImagesIndexToVideoOutput"] | components["schemas"]["IntegerCollectionOutput"] | components["schemas"]["IntegerGeneratorOutput"] | components["schemas"]["IntegerOutput"] | components["schemas"]["IterateInvocationOutput"] | components["schemas"]["JsonListStringsOutput"] | components["schemas"]["LatentsCollectionOutput"] | components["schemas"]["LatentsMetaOutput"] | components["schemas"]["LatentsOutput"] | components["schemas"]["LoRACollectionFromPathOutput"] | components["schemas"]["LoRACollectionOutput"] | components["schemas"]["LoRACollectionToggleOutput"] | components["schemas"]["LoRALoaderOutput"] | components["schemas"]["LoRANameGrabberOutput"] | components["schemas"]["LoRASelectorOutput"] | components["schemas"]["LoRAToggleOutput"] | components["schemas"]["LoadAllTextFilesInFolderOutput"] | components["schemas"]["LoadTextFileToStringOutput"] | components["schemas"]["LookupLoRACollectionTriggersOutput"] | components["schemas"]["LookupLoRATriggersOutput"] | components["schemas"]["LookupTableOutput"] | components["schemas"]["MDControlListOutput"] | components["schemas"]["MDIPAdapterListOutput"] | components["schemas"]["MDT2IAdapterListOutput"] | components["schemas"]["MaskOutput"] | components["schemas"]["MathEvalOutput"] | components["schemas"]["MergeLoRACollectionsOutput"] | components["schemas"]["MergeStringCollectionsOutput"] | components["schemas"]["MetadataItemOutput"] | components["schemas"]["MetadataOutput"] | components["schemas"]["MetadataToLorasCollectionOutput"] | components["schemas"]["MetadataToModelOutput"] | components["schemas"]["MetadataToSDXLModelOutput"] | components["schemas"]["ModelIdentifierOutput"] | components["schemas"]["ModelLoaderOutput"] | components["schemas"]["ModelNameGrabberOutput"] | components["schemas"]["ModelToggleOutput"] | components["schemas"]["NightmareOutput"] | components["schemas"]["NoiseOutput"] | components["schemas"]["PBRMapsOutput"] | components["schemas"]["PTFieldsCollectOutput"] | components["schemas"]["PTFieldsExpandOutput"] | components["schemas"]["PairTileImageOutput"] | components["schemas"]["PaletteOutput"] | components["schemas"]["PrintStringToConsoleOutput"] | components["schemas"]["PromptStrengthOutput"] | components["schemas"]["PromptTemplateOutput"] | components["schemas"]["PromptsToFileInvocationOutput"] | components["schemas"]["PrunedPromptOutput"] | components["schemas"]["Qwen3PromptProOutput"] | components["schemas"]["RGBSplitOutput"] | components["schemas"]["RandomImageSizeOutput"] | components["schemas"]["RandomLoRAMixerOutput"] | components["schemas"]["ReapplyLoRAWeightOutput"] | components["schemas"]["RetrieveFluxConditioningMultiOutput"] | components["schemas"]["SD3ConditioningOutput"] | components["schemas"]["SDXLLoRALoaderOutput"] | components["schemas"]["SDXLModelLoaderOutput"] | components["schemas"]["SDXLRefinerModelLoaderOutput"] | components["schemas"]["SchedulerOutput"] | components["schemas"]["Sd3ModelLoaderOutput"] | components["schemas"]["SeamlessModeOutput"] | components["schemas"]["ShadowsHighlightsMidtonesMasksOutput"] | components["schemas"]["String2Output"] | components["schemas"]["StringCollectionJoinerOutput"] | components["schemas"]["StringCollectionOutput"] | components["schemas"]["StringGeneratorOutput"] | components["schemas"]["StringOutput"] | components["schemas"]["StringPosNegOutput"] | components["schemas"]["StringToLoraOutput"] | components["schemas"]["StringToMainModelOutput"] | components["schemas"]["StringToModelOutput"] | components["schemas"]["StringToSDXLModelOutput"] | components["schemas"]["T2IAdapterListOutput"] | components["schemas"]["T2IAdapterOutput"] | components["schemas"]["ThresholdingOutput"] | components["schemas"]["TileSizeOutput"] | components["schemas"]["TileToPropertiesOutput"] | components["schemas"]["TilesOutput"] | components["schemas"]["TraceryOutput"] | components["schemas"]["UNetOutput"] | components["schemas"]["VAEOutput"] | components["schemas"]["WeightedStringOutput"] | components["schemas"]["XYExpandOutput"] | components["schemas"]["XYImageExpandOutput"] | components["schemas"]["XYProductOutput"] | components["schemas"]["ZImageConditioningOutput"] | components["schemas"]["ZImageControlOutput"] | components["schemas"]["ZImageLoRALoaderOutput"] | components["schemas"]["ZImageModelLoaderOutput"]; + }; + /** + * InvocationErrorEvent + * @description Event model for invocation_error + */ + InvocationErrorEvent: { + /** + * Timestamp + * @description The timestamp of the event + */ + timestamp: number; + /** + * Queue Id + * @description The ID of the queue + */ + queue_id: string; + /** + * Item Id + * @description The ID of the queue item + */ + item_id: number; + /** + * Batch Id + * @description The ID of the queue batch + */ + batch_id: string; + /** + * Origin + * @description The origin of the queue item + * @default null + */ + origin: string | null; + /** + * Destination + * @description The destination of the queue item + * @default null + */ + destination: string | null; + /** + * User Id + * @description The ID of the user who created the queue item + * @default system + */ + user_id: string; + /** + * Session Id + * @description The ID of the session (aka graph execution state) + */ + session_id: string; + /** + * Invocation + * @description The ID of the invocation + */ + invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + /** + * Invocation Source Id + * @description The ID of the prepared invocation's source node + */ + invocation_source_id: string; + /** + * Error Type + * @description The error type + */ + error_type: string; + /** + * Error Message + * @description The error message + */ + error_message: string; + /** + * Error Traceback + * @description The error traceback + */ + error_traceback: string; + }; + InvocationOutputMap: { + Advanced_Text_Font_to_Image: components["schemas"]["ImageOutput"]; + I2AA_AnyFont: components["schemas"]["ImageOutput"]; + Image_to_ASCII_Art_Image: components["schemas"]["ImageOutput"]; + Image_to_Unicode_Art: components["schemas"]["ImageOutput"]; + Retrieve_Board_Images: components["schemas"]["ImageCollectionOutput"]; + Text_Font_to_Image: components["schemas"]["ImageOutput"]; + add: components["schemas"]["IntegerOutput"]; + adv_autostereogram: components["schemas"]["ImageOutput"]; + alpha_mask_to_tensor: components["schemas"]["MaskOutput"]; + anamorphic_streaks: components["schemas"]["ImageOutput"]; + apply_mask_to_image: components["schemas"]["ImageOutput"]; + apply_tensor_mask_to_image: components["schemas"]["ImageOutput"]; + autostereogram: components["schemas"]["ImageOutput"]; + average_images: components["schemas"]["ImageOutput"]; + blank_image: components["schemas"]["ImageOutput"]; + bool_collection_index: components["schemas"]["BooleanOutput"]; + bool_collection_toggle: components["schemas"]["BooleanCollectionOutput"]; + bool_toggle: components["schemas"]["BooleanOutput"]; + boolean: components["schemas"]["BooleanOutput"]; + boolean_collection: components["schemas"]["BooleanCollectionOutput"]; + boolean_collection_linked: components["schemas"]["BooleanCollectionOutput"]; + bounding_box: components["schemas"]["BoundingBoxOutput"]; + calculate_image_tiles: components["schemas"]["CalculateImageTilesOutput"]; + calculate_image_tiles_even_split: components["schemas"]["CalculateImageTilesOutput"]; + calculate_image_tiles_min_overlap: components["schemas"]["CalculateImageTilesOutput"]; + canny_edge_detection: components["schemas"]["ImageOutput"]; + canvas_paste_back: components["schemas"]["ImageOutput"]; + canvas_v2_mask_and_crop: components["schemas"]["ImageOutput"]; + chroma_noise: components["schemas"]["ImageOutput"]; + chromatic_aberration: components["schemas"]["ImageOutput"]; + clip_skip: components["schemas"]["CLIPSkipInvocationOutput"]; + clipseg_mask_hierarchy: components["schemas"]["ClipsegMaskHierarchyOutput"]; + cmyk_halftone: components["schemas"]["ImageOutput"]; + cmyk_merge: components["schemas"]["ImageOutput"]; + cmyk_separation: components["schemas"]["CMYKSeparationOutput"]; + cmyk_split: components["schemas"]["CMYKSplitOutput"]; + cogview4_denoise: components["schemas"]["LatentsOutput"]; + cogview4_i2l: components["schemas"]["LatentsOutput"]; + cogview4_l2i: components["schemas"]["ImageOutput"]; + cogview4_model_loader: components["schemas"]["CogView4ModelLoaderOutput"]; + cogview4_text_encoder: components["schemas"]["CogView4ConditioningOutput"]; + collect: components["schemas"]["CollectInvocationOutput"]; + collection_count: components["schemas"]["CollectionCountOutput"]; + collection_index: components["schemas"]["CollectionIndexOutput"]; + collection_join: components["schemas"]["CollectionJoinOutput"]; + collection_reverse: components["schemas"]["CollectionReverseOutput"]; + collection_slice: components["schemas"]["CollectionSliceOutput"]; + collection_sort: components["schemas"]["CollectionSortOutput"]; + collection_unique: components["schemas"]["CollectionUniqueOutput"]; + color: components["schemas"]["ColorOutput"]; + color_cast_correction: components["schemas"]["ImageOutput"]; + color_correct: components["schemas"]["ImageOutput"]; + color_map: components["schemas"]["ImageOutput"]; + compare_floats_invocation: components["schemas"]["BooleanOutput"]; + compare_ints_invocation: components["schemas"]["BooleanOutput"]; + compare_strings_invocation: components["schemas"]["BooleanOutput"]; + compel: components["schemas"]["ConditioningOutput"]; + conditioning: components["schemas"]["ConditioningOutput"]; + conditioning_collection: components["schemas"]["ConditioningCollectionOutput"]; + conditioning_collection_linked: components["schemas"]["ConditioningCollectionOutput"]; + conditioning_collection_toggle: components["schemas"]["ConditioningCollectionOutput"]; + conditioning_toggle: components["schemas"]["ConditioningOutput"]; + content_shuffle: components["schemas"]["ImageOutput"]; + controlnet: components["schemas"]["ControlOutput"]; + "controlnet-linked": components["schemas"]["ControlListOutput"]; + core_metadata: components["schemas"]["MetadataOutput"]; + create_denoise_mask: components["schemas"]["DenoiseMaskOutput"]; + create_gradient_mask: components["schemas"]["GradientMaskOutput"]; + crop_image_to_bounding_box: components["schemas"]["ImageOutput"]; + crop_latents: components["schemas"]["LatentsOutput"]; + crossover_prompt: components["schemas"]["HalvedPromptOutput"]; + csv_to_index_string: components["schemas"]["StringOutput"]; + csv_to_strings: components["schemas"]["StringCollectionOutput"]; + cv_extrude_depth: components["schemas"]["ImageOutput"]; + cv_inpaint: components["schemas"]["ImageOutput"]; + decode_watermark: components["schemas"]["StringOutput"]; + default_xy_tile_generator: components["schemas"]["TilesOutput"]; + denoise_latents: components["schemas"]["LatentsOutput"]; + denoise_latents_meta: components["schemas"]["LatentsMetaOutput"]; + depth_anything_depth_estimation: components["schemas"]["ImageOutput"]; + div: components["schemas"]["IntegerOutput"]; + dw_openpose_detection: components["schemas"]["ImageOutput"]; + dynamic_prompt: components["schemas"]["StringCollectionOutput"]; + enhance_detail: components["schemas"]["ImageOutput"]; + esrgan: components["schemas"]["ImageOutput"]; + even_split_xy_tile_generator: components["schemas"]["TilesOutput"]; + expand_mask_with_fade: components["schemas"]["ImageOutput"]; + extract_image_collection_metadata_boolean: components["schemas"]["BooleanCollectionOutput"]; + extract_image_collection_metadata_float: components["schemas"]["FloatCollectionOutput"]; + extract_image_collection_metadata_integer: components["schemas"]["IntegerCollectionOutput"]; + extract_image_collection_metadata_string: components["schemas"]["StringCollectionOutput"]; + face_identifier: components["schemas"]["ImageOutput"]; + face_mask_detection: components["schemas"]["FaceMaskOutput"]; + face_off: components["schemas"]["FaceOffOutput"]; + film_grain: components["schemas"]["ImageOutput"]; + flatten_histogram_mono: components["schemas"]["ImageOutput"]; + float: components["schemas"]["FloatOutput"]; + float_batch: components["schemas"]["FloatOutput"]; + float_collection: components["schemas"]["FloatCollectionOutput"]; + float_collection_index: components["schemas"]["FloatOutput"]; + float_collection_linked: components["schemas"]["FloatCollectionOutput"]; + float_collection_toggle: components["schemas"]["FloatCollectionOutput"]; + float_generator: components["schemas"]["FloatGeneratorOutput"]; + float_math: components["schemas"]["FloatOutput"]; + float_range: components["schemas"]["FloatCollectionOutput"]; + float_to_int: components["schemas"]["IntegerOutput"]; + float_toggle: components["schemas"]["FloatOutput"]; + floats_to_strings: components["schemas"]["StringCollectionOutput"]; + flux2_denoise: components["schemas"]["LatentsOutput"]; + flux2_klein_lora_collection_loader: components["schemas"]["Flux2KleinLoRALoaderOutput"]; + flux2_klein_lora_loader: components["schemas"]["Flux2KleinLoRALoaderOutput"]; + flux2_klein_model_loader: components["schemas"]["Flux2KleinModelLoaderOutput"]; + flux2_klein_text_encoder: components["schemas"]["FluxConditioningOutput"]; + flux2_vae_decode: components["schemas"]["ImageOutput"]; + flux2_vae_encode: components["schemas"]["LatentsOutput"]; + flux_conditioning_blend: components["schemas"]["FluxConditioningBlendOutput"]; + flux_conditioning_collection: components["schemas"]["FluxConditioningCollectionOutput"]; + flux_conditioning_collection_join: components["schemas"]["FluxConditioningCollectionOutput"]; + flux_conditioning_collection_toggle: components["schemas"]["FluxConditioningCollectionOutput"]; + flux_conditioning_concatenate: components["schemas"]["FluxConditioningOutput"]; + flux_conditioning_delta_augmentation: components["schemas"]["FluxConditioningDeltaAndAugmentedOutput"]; + flux_conditioning_index: components["schemas"]["FluxConditioningOutput"]; + flux_conditioning_list: components["schemas"]["FluxConditioningListOutput"]; + flux_conditioning_math: components["schemas"]["FluxConditioningOutput"]; + flux_conditioning_toggle: components["schemas"]["FluxConditioningOutput"]; + flux_control_lora_loader: components["schemas"]["FluxControlLoRALoaderOutput"]; + flux_controlnet: components["schemas"]["FluxControlNetOutput"]; + flux_controlnet_collection: components["schemas"]["FluxControlNetCollectionOutput"]; + flux_controlnet_collection_join: components["schemas"]["FluxControlNetCollectionOutput"]; + flux_controlnet_index: components["schemas"]["FluxControlNetOutput"]; + flux_controlnet_linked: components["schemas"]["FluxControlNetListOutput"]; + flux_denoise: components["schemas"]["LatentsOutput"]; + flux_denoise_meta: components["schemas"]["LatentsMetaOutput"]; + flux_fill: components["schemas"]["FluxFillOutput"]; + flux_ideal_size: components["schemas"]["FluxIdealSizeOutput"]; + flux_ip_adapter: components["schemas"]["IPAdapterOutput"]; + flux_kontext: components["schemas"]["FluxKontextOutput"]; + flux_kontext_ideal_size: components["schemas"]["FluxIdealSizeOutput"]; + flux_kontext_image_prep: components["schemas"]["ImageOutput"]; + flux_lora_collection_loader: components["schemas"]["FluxLoRALoaderOutput"]; + flux_lora_loader: components["schemas"]["FluxLoRALoaderOutput"]; + flux_model_loader: components["schemas"]["FluxModelLoaderOutput"]; + flux_model_loader_input: components["schemas"]["FluxModelLoaderOutput"]; + flux_model_to_string: components["schemas"]["StringOutput"]; + flux_redux: components["schemas"]["FluxReduxOutput"]; + flux_redux_collection: components["schemas"]["FluxReduxCollectionOutput"]; + flux_redux_collection_join: components["schemas"]["FluxReduxCollectionOutput"]; + flux_redux_conditioning_math: components["schemas"]["FluxReduxOutput"]; + flux_redux_downsampling: components["schemas"]["FluxReduxOutput"]; + flux_redux_index: components["schemas"]["FluxReduxOutput"]; + flux_redux_linked: components["schemas"]["FluxReduxListOutput"]; + flux_redux_rescale_conditioning: components["schemas"]["FluxReduxOutput"]; + flux_rescale_conditioning: components["schemas"]["FluxConditioningOutput"]; + flux_scale_conditioning: components["schemas"]["FluxConditioningOutput"]; + flux_scale_prompt_section: components["schemas"]["FluxConditioningOutput"]; + flux_scale_redux_conditioning: components["schemas"]["FluxReduxOutput"]; + flux_text_encoder: components["schemas"]["FluxConditioningOutput"]; + flux_text_encoder_linked: components["schemas"]["FluxTextEncoderListOutput"]; + flux_vae_decode: components["schemas"]["ImageOutput"]; + flux_vae_encode: components["schemas"]["LatentsOutput"]; + flux_weighted_prompt: components["schemas"]["FluxWeightedPromptOutput"]; + freeu: components["schemas"]["UNetOutput"]; + frequency_blend_latents: components["schemas"]["LatentsOutput"]; + frequency_match_latents: components["schemas"]["LatentsOutput"]; + generate_evolutionary_prompts: components["schemas"]["EvolutionaryPromptListOutput"]; + get_image_mask_bounding_box: components["schemas"]["BoundingBoxOutput"]; + get_palette: components["schemas"]["ImageOutput"]; + get_palette_adv: components["schemas"]["ImageOutput"]; + get_source_framerate: components["schemas"]["FloatOutput"]; + get_total_frames: components["schemas"]["IntegerOutput"]; + grounding_dino: components["schemas"]["BoundingBoxCollectionOutput"]; + halftone: components["schemas"]["ImageOutput"]; + hed_edge_detection: components["schemas"]["ImageOutput"]; + heuristic_resize: components["schemas"]["ImageOutput"]; + i2l: components["schemas"]["LatentsOutput"]; + ideal_size: components["schemas"]["IdealSizeOutput"]; + image: components["schemas"]["ImageOutput"]; + image_batch: components["schemas"]["ImageOutput"]; + image_collection: components["schemas"]["ImageCollectionOutput"]; + image_collection_index: components["schemas"]["ImageOutput"]; + image_collection_linked: components["schemas"]["ImageCollectionOutput"]; + image_collection_toggle: components["schemas"]["ImageCollectionOutput"]; + image_generator: components["schemas"]["ImageGeneratorOutput"]; + image_index_collect: components["schemas"]["ImageIndexCollectOutput"]; + image_mask_to_tensor: components["schemas"]["MaskOutput"]; + image_panel_layout: components["schemas"]["ImagePanelCoordinateOutput"]; + image_som: components["schemas"]["ImageSOMOutput"]; + image_to_image_name: components["schemas"]["ImageToImageNameOutput"]; + image_to_xy_image_collection: components["schemas"]["StringCollectionOutput"]; + image_to_xy_image_tiles: components["schemas"]["ImageToXYImageTilesOutput"]; + image_toggle: components["schemas"]["ImageOutput"]; + images_index_to_video: components["schemas"]["ImagesIndexToVideoOutput"]; + images_to_grids: components["schemas"]["ImageCollectionOutput"]; + img_blur: components["schemas"]["ImageOutput"]; + img_chan: components["schemas"]["ImageOutput"]; + img_channel_multiply: components["schemas"]["ImageOutput"]; + img_channel_offset: components["schemas"]["ImageOutput"]; + img_conv: components["schemas"]["ImageOutput"]; + img_crop: components["schemas"]["ImageOutput"]; + img_hue_adjust: components["schemas"]["ImageOutput"]; + img_ilerp: components["schemas"]["ImageOutput"]; + img_lerp: components["schemas"]["ImageOutput"]; + img_mul: components["schemas"]["ImageOutput"]; + img_noise: components["schemas"]["ImageOutput"]; + img_nsfw: components["schemas"]["ImageOutput"]; img_pad_crop: components["schemas"]["ImageOutput"]; img_paste: components["schemas"]["ImageOutput"]; img_resize: components["schemas"]["ImageOutput"]; img_scale: components["schemas"]["ImageOutput"]; img_watermark: components["schemas"]["ImageOutput"]; + imgs2mask_clipseg: components["schemas"]["ImageOutput"]; infill_cv2: components["schemas"]["ImageOutput"]; infill_lama: components["schemas"]["ImageOutput"]; infill_patchmatch: components["schemas"]["ImageOutput"]; infill_rgba: components["schemas"]["ImageOutput"]; infill_tile: components["schemas"]["ImageOutput"]; + info_grabber_unet_invocation: components["schemas"]["StringOutput"]; + int_collection_toggle: components["schemas"]["IntegerCollectionOutput"]; + int_toggle: components["schemas"]["IntegerOutput"]; integer: components["schemas"]["IntegerOutput"]; integer_batch: components["schemas"]["IntegerOutput"]; integer_collection: components["schemas"]["IntegerCollectionOutput"]; + integer_collection_index: components["schemas"]["IntegerOutput"]; + integer_collection_linked: components["schemas"]["IntegerCollectionOutput"]; integer_generator: components["schemas"]["IntegerGeneratorOutput"]; integer_math: components["schemas"]["IntegerOutput"]; + ints_to_strings: components["schemas"]["StringCollectionOutput"]; invert_tensor_mask: components["schemas"]["MaskOutput"]; invokeai_ealightness: components["schemas"]["ImageOutput"]; invokeai_img_blend: components["schemas"]["ImageOutput"]; @@ -14182,25 +21353,70 @@ export type components = { invokeai_img_hue_adjust_plus: components["schemas"]["ImageOutput"]; invokeai_img_val_thresholds: components["schemas"]["ImageOutput"]; ip_adapter: components["schemas"]["IPAdapterOutput"]; + ip_adapter_linked: components["schemas"]["IPAdapterListOutput"]; iterate: components["schemas"]["IterateInvocationOutput"]; l2i: components["schemas"]["ImageOutput"]; + latent_average: components["schemas"]["LatentsOutput"]; + latent_band_pass_filter: components["schemas"]["LatentsOutput"]; + latent_blend_linear: components["schemas"]["LatentsOutput"]; + latent_channels_to_grid: components["schemas"]["ImageCollectionOutput"]; + latent_combine: components["schemas"]["LatentsOutput"]; + latent_dtype_convert: components["schemas"]["LatentsOutput"]; + latent_high_pass_filter: components["schemas"]["LatentsOutput"]; + latent_low_pass_filter: components["schemas"]["LatentsOutput"]; + latent_match: components["schemas"]["LatentsOutput"]; + latent_modify_channels: components["schemas"]["LatentsOutput"]; + latent_normalize_range: components["schemas"]["LatentsOutput"]; + latent_normalize_std_dev: components["schemas"]["LatentsOutput"]; + latent_normalize_std_dev_range: components["schemas"]["LatentsOutput"]; + latent_plot: components["schemas"]["ImageCollectionOutput"]; + latent_som: components["schemas"]["LatentsOutput"]; + latent_white_noise: components["schemas"]["LatentsOutput"]; latents: components["schemas"]["LatentsOutput"]; latents_collection: components["schemas"]["LatentsCollectionOutput"]; + latents_collection_index: components["schemas"]["LatentsOutput"]; + latents_collection_linked: components["schemas"]["LatentsCollectionOutput"]; lblend: components["schemas"]["LatentsOutput"]; + lcrop: components["schemas"]["LatentsOutput"]; + lens_aperture_generator: components["schemas"]["ImageOutput"]; + lens_blur: components["schemas"]["ImageOutput"]; + lens_vignette: components["schemas"]["ImageOutput"]; lineart_anime_edge_detection: components["schemas"]["ImageOutput"]; lineart_edge_detection: components["schemas"]["ImageOutput"]; llava_onevision_vllm: components["schemas"]["StringOutput"]; + load_all_text_files_in_folder_output: components["schemas"]["LoadAllTextFilesInFolderOutput"]; + load_aperture_image: components["schemas"]["ImageOutput"]; + load_text_file_to_string_invocation: components["schemas"]["LoadTextFileToStringOutput"]; + load_video_frame: components["schemas"]["ImageOutput"]; + lookup_from_prompt: components["schemas"]["LookupTableOutput"]; + lookup_lora_collection_triggers_invocation: components["schemas"]["LookupLoRATriggersOutput"]; + lookup_lora_triggers_invocation: components["schemas"]["LookupLoRATriggersOutput"]; + lookup_table_from_file: components["schemas"]["LookupTableOutput"]; + lora_collection: components["schemas"]["LoRACollectionOutput"]; + lora_collection_from_path_invocation: components["schemas"]["LoRACollectionFromPathOutput"]; + lora_collection_linked: components["schemas"]["LoRACollectionOutput"]; lora_collection_loader: components["schemas"]["LoRALoaderOutput"]; + lora_collection_toggle: components["schemas"]["LoRACollectionToggleOutput"]; lora_loader: components["schemas"]["LoRALoaderOutput"]; + lora_name_grabber_invocation: components["schemas"]["LoRANameGrabberOutput"]; lora_selector: components["schemas"]["LoRASelectorOutput"]; + lora_to_string: components["schemas"]["StringOutput"]; + lora_toggle: components["schemas"]["LoRAToggleOutput"]; lresize: components["schemas"]["LatentsOutput"]; lscale: components["schemas"]["LatentsOutput"]; main_model_loader: components["schemas"]["ModelLoaderOutput"]; + main_model_loader_input: components["schemas"]["ModelLoaderOutput"]; + main_model_to_string: components["schemas"]["StringOutput"]; mask_combine: components["schemas"]["ImageOutput"]; mask_edge: components["schemas"]["ImageOutput"]; mask_from_id: components["schemas"]["ImageOutput"]; + match_histogram: components["schemas"]["ImageOutput"]; + match_histogram_lab: components["schemas"]["ImageOutput"]; + math_eval: components["schemas"]["MathEvalOutput"]; mediapipe_face_detection: components["schemas"]["ImageOutput"]; + merge_lora_collections_invocation: components["schemas"]["MergeLoRACollectionsOutput"]; merge_metadata: components["schemas"]["MetadataOutput"]; + merge_string_collections_invocation: components["schemas"]["MergeStringCollectionsOutput"]; merge_tiles_to_image: components["schemas"]["ImageOutput"]; metadata: components["schemas"]["MetadataOutput"]; metadata_field_extractor: components["schemas"]["StringOutput"]; @@ -14225,59 +21441,142 @@ export type components = { metadata_to_string_collection: components["schemas"]["StringCollectionOutput"]; metadata_to_t2i_adapters: components["schemas"]["MDT2IAdapterListOutput"]; metadata_to_vae: components["schemas"]["VAEOutput"]; + minimum_overlap_xy_tile_generator: components["schemas"]["TilesOutput"]; mlsd_detection: components["schemas"]["ImageOutput"]; model_identifier: components["schemas"]["ModelIdentifierOutput"]; + model_name_grabber_invocation: components["schemas"]["ModelNameGrabberOutput"]; + model_name_to_model: components["schemas"]["StringToModelOutput"]; + model_to_string: components["schemas"]["StringOutput"]; + model_toggle: components["schemas"]["ModelToggleOutput"]; + monochrome_film_grain: components["schemas"]["ImageOutput"]; mul: components["schemas"]["IntegerOutput"]; + nightmare_promptgen: components["schemas"]["NightmareOutput"]; noise: components["schemas"]["NoiseOutput"]; + noise_add_flux: components["schemas"]["LatentsOutput"]; + noise_coordinated: components["schemas"]["NoiseOutput"]; + noise_coordinated_flux: components["schemas"]["LatentsOutput"]; + noise_spectral: components["schemas"]["NoiseOutput"]; + noiseimg_2d: components["schemas"]["ImageOutput"]; normal_map: components["schemas"]["ImageOutput"]; + octree_quantizer: components["schemas"]["ImageOutput"]; + offset_image: components["schemas"]["ImageOutput"]; + offset_latents: components["schemas"]["LatentsOutput"]; + optimized_tile_size_from_area: components["schemas"]["TileSizeOutput"]; pair_tile_image: components["schemas"]["PairTileImageOutput"]; + parse_weighted_string: components["schemas"]["WeightedStringOutput"]; paste_image_into_bounding_box: components["schemas"]["ImageOutput"]; pbr_maps: components["schemas"]["PBRMapsOutput"]; + percent_to_float: components["schemas"]["FloatOutput"]; pidi_edge_detection: components["schemas"]["ImageOutput"]; + "pixel-art": components["schemas"]["ImageOutput"]; + pixelize: components["schemas"]["ImageOutput"]; + print_string_to_console_invocation: components["schemas"]["PrintStringToConsoleOutput"]; + prompt_auto_and: components["schemas"]["StringOutput"]; prompt_from_file: components["schemas"]["StringCollectionOutput"]; + prompt_from_lookup_table: components["schemas"]["HalvedPromptOutput"]; + prompt_strength: components["schemas"]["PromptStrengthOutput"]; + prompt_strengths_combine: components["schemas"]["StringOutput"]; prompt_template: components["schemas"]["PromptTemplateOutput"]; + prompt_to_file: components["schemas"]["PromptsToFileInvocationOutput"]; + prune_prompt: components["schemas"]["PrunedPromptOutput"]; + pt_fields_collect: components["schemas"]["PTFieldsCollectOutput"]; + pt_fields_expand: components["schemas"]["PTFieldsExpandOutput"]; + quote_escaper: components["schemas"]["EscapedOutput"]; + qwen3_prompt_pro: components["schemas"]["Qwen3PromptProOutput"]; rand_float: components["schemas"]["FloatOutput"]; rand_int: components["schemas"]["IntegerOutput"]; + random_image_size_invocation: components["schemas"]["RandomImageSizeOutput"]; + random_lora_mixer_invocation: components["schemas"]["RandomLoRAMixerOutput"]; random_range: components["schemas"]["IntegerCollectionOutput"]; range: components["schemas"]["IntegerCollectionOutput"]; range_of_size: components["schemas"]["IntegerCollectionOutput"]; + reapply_lora_weight_invocation: components["schemas"]["ReapplyLoRAWeightOutput"]; rectangle_mask: components["schemas"]["MaskOutput"]; + retrieve_flux_conditioning: components["schemas"]["RetrieveFluxConditioningMultiOutput"]; + retro_bitize: components["schemas"]["ImageOutput"]; + retro_crt_curvature: components["schemas"]["ImageOutput"]; + retro_palettize: components["schemas"]["ImageOutput"]; + retro_palettize_adv: components["schemas"]["ImageOutput"]; + retro_pixelize: components["schemas"]["ImageOutput"]; + retro_quantize: components["schemas"]["ImageOutput"]; + retro_scanlines_simple: components["schemas"]["ImageOutput"]; + rgb_merge: components["schemas"]["ImageOutput"]; + rgb_split: components["schemas"]["RGBSplitOutput"]; + rotate_image: components["schemas"]["ImageOutput"]; round_float: components["schemas"]["FloatOutput"]; save_image: components["schemas"]["ImageOutput"]; scheduler: components["schemas"]["SchedulerOutput"]; + scheduler_to_string: components["schemas"]["StringOutput"]; + scheduler_toggle: components["schemas"]["SchedulerOutput"]; sd3_denoise: components["schemas"]["LatentsOutput"]; sd3_i2l: components["schemas"]["LatentsOutput"]; sd3_l2i: components["schemas"]["ImageOutput"]; sd3_model_loader: components["schemas"]["Sd3ModelLoaderOutput"]; + sd3_model_loader_input: components["schemas"]["Sd3ModelLoaderOutput"]; + sd3_model_to_string: components["schemas"]["StringOutput"]; sd3_text_encoder: components["schemas"]["SD3ConditioningOutput"]; sdxl_compel_prompt: components["schemas"]["ConditioningOutput"]; sdxl_lora_collection_loader: components["schemas"]["SDXLLoRALoaderOutput"]; sdxl_lora_loader: components["schemas"]["SDXLLoRALoaderOutput"]; + sdxl_main_model_toggle: components["schemas"]["SDXLModelLoaderOutput"]; sdxl_model_loader: components["schemas"]["SDXLModelLoaderOutput"]; + sdxl_model_loader_input: components["schemas"]["SDXLModelLoaderOutput"]; + sdxl_model_to_string: components["schemas"]["StringOutput"]; sdxl_refiner_compel_prompt: components["schemas"]["ConditioningOutput"]; sdxl_refiner_model_loader: components["schemas"]["SDXLRefinerModelLoaderOutput"]; seamless: components["schemas"]["SeamlessModeOutput"]; segment_anything: components["schemas"]["MaskOutput"]; + separate_prompt_and_seed_vector: components["schemas"]["JsonListStringsOutput"]; + shmmask: components["schemas"]["ShadowsHighlightsMidtonesMasksOutput"]; show_image: components["schemas"]["ImageOutput"]; spandrel_image_to_image: components["schemas"]["ImageOutput"]; spandrel_image_to_image_autoscale: components["schemas"]["ImageOutput"]; + spherical_distortion: components["schemas"]["ImageOutput"]; + store_flux_conditioning: components["schemas"]["FluxConditioningStoreOutput"]; string: components["schemas"]["StringOutput"]; string_batch: components["schemas"]["StringOutput"]; string_collection: components["schemas"]["StringCollectionOutput"]; + string_collection_index: components["schemas"]["StringOutput"]; + string_collection_joiner_invocation: components["schemas"]["StringCollectionJoinerOutput"]; + string_collection_linked: components["schemas"]["StringCollectionOutput"]; + string_collection_toggle: components["schemas"]["StringCollectionOutput"]; string_generator: components["schemas"]["StringGeneratorOutput"]; string_join: components["schemas"]["StringOutput"]; string_join_three: components["schemas"]["StringOutput"]; string_replace: components["schemas"]["StringOutput"]; string_split: components["schemas"]["String2Output"]; string_split_neg: components["schemas"]["StringPosNegOutput"]; + string_to_collection_splitter_invocation: components["schemas"]["StringCollectionOutput"]; + string_to_float: components["schemas"]["FloatOutput"]; + string_to_int: components["schemas"]["IntegerOutput"]; + string_to_lora: components["schemas"]["StringToLoraOutput"]; + string_to_main_model: components["schemas"]["StringToMainModelOutput"]; + string_to_model: components["schemas"]["StringToModelOutput"]; + string_to_scheduler: components["schemas"]["SchedulerOutput"]; + string_to_sdxl_model: components["schemas"]["StringToSDXLModelOutput"]; + string_toggle: components["schemas"]["StringOutput"]; + strings_to_csv: components["schemas"]["StringOutput"]; sub: components["schemas"]["IntegerOutput"]; t2i_adapter: components["schemas"]["T2IAdapterOutput"]; + t2i_adapter_linked: components["schemas"]["T2IAdapterListOutput"]; tensor_mask_to_image: components["schemas"]["ImageOutput"]; + text_mask: components["schemas"]["ImageOutput"]; + thresholding: components["schemas"]["ThresholdingOutput"]; tile_to_properties: components["schemas"]["TileToPropertiesOutput"]; tiled_multi_diffusion_denoise_latents: components["schemas"]["LatentsOutput"]; tomask: components["schemas"]["ImageOutput"]; + tracery_invocation: components["schemas"]["TraceryOutput"]; + txt2mask_clipseg: components["schemas"]["ImageOutput"]; + txt2mask_clipseg_adv: components["schemas"]["ImageOutput"]; unsharp_mask: components["schemas"]["ImageOutput"]; vae_loader: components["schemas"]["VAEOutput"]; + xy_expand: components["schemas"]["XYExpandOutput"]; + xy_image_collect: components["schemas"]["StringOutput"]; + xy_image_expand: components["schemas"]["XYImageExpandOutput"]; + xy_image_tiles_to_image: components["schemas"]["ImageOutput"]; + xy_images_to_grid: components["schemas"]["ImageOutput"]; + xy_product: components["schemas"]["XYProductOutput"]; + xy_product_csv: components["schemas"]["XYProductOutput"]; z_image_control: components["schemas"]["ZImageControlOutput"]; z_image_denoise: components["schemas"]["LatentsOutput"]; z_image_denoise_meta: components["schemas"]["LatentsMetaOutput"]; @@ -14290,905 +21589,7055 @@ export type components = { z_image_text_encoder: components["schemas"]["ZImageConditioningOutput"]; }; /** - * InvocationProgressEvent - * @description Event model for invocation_progress + * InvocationProgressEvent + * @description Event model for invocation_progress + */ + InvocationProgressEvent: { + /** + * Timestamp + * @description The timestamp of the event + */ + timestamp: number; + /** + * Queue Id + * @description The ID of the queue + */ + queue_id: string; + /** + * Item Id + * @description The ID of the queue item + */ + item_id: number; + /** + * Batch Id + * @description The ID of the queue batch + */ + batch_id: string; + /** + * Origin + * @description The origin of the queue item + * @default null + */ + origin: string | null; + /** + * Destination + * @description The destination of the queue item + * @default null + */ + destination: string | null; + /** + * User Id + * @description The ID of the user who created the queue item + * @default system + */ + user_id: string; + /** + * Session Id + * @description The ID of the session (aka graph execution state) + */ + session_id: string; + /** + * Invocation + * @description The ID of the invocation + */ + invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + /** + * Invocation Source Id + * @description The ID of the prepared invocation's source node + */ + invocation_source_id: string; + /** + * Message + * @description A message to display + */ + message: string; + /** + * Percentage + * @description The percentage of the progress (omit to indicate indeterminate progress) + * @default null + */ + percentage: number | null; + /** + * @description An image representing the current state of the progress + * @default null + */ + image: components["schemas"]["ProgressImage"] | null; + }; + /** + * InvocationStartedEvent + * @description Event model for invocation_started + */ + InvocationStartedEvent: { + /** + * Timestamp + * @description The timestamp of the event + */ + timestamp: number; + /** + * Queue Id + * @description The ID of the queue + */ + queue_id: string; + /** + * Item Id + * @description The ID of the queue item + */ + item_id: number; + /** + * Batch Id + * @description The ID of the queue batch + */ + batch_id: string; + /** + * Origin + * @description The origin of the queue item + * @default null + */ + origin: string | null; + /** + * Destination + * @description The destination of the queue item + * @default null + */ + destination: string | null; + /** + * User Id + * @description The ID of the user who created the queue item + * @default system + */ + user_id: string; + /** + * Session Id + * @description The ID of the session (aka graph execution state) + */ + session_id: string; + /** + * Invocation + * @description The ID of the invocation + */ + invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + /** + * Invocation Source Id + * @description The ID of the prepared invocation's source node + */ + invocation_source_id: string; + }; + /** + * InvokeAIAppConfig + * @description Invoke's global app configuration. + * + * Typically, you won't need to interact with this class directly. Instead, use the `get_config` function from `invokeai.app.services.config` to get a singleton config object. + * + * Attributes: + * host: IP address to bind to. Use `0.0.0.0` to serve to your local network. + * port: Port to bind to. + * allow_origins: Allowed CORS origins. + * allow_credentials: Allow CORS credentials. + * allow_methods: Methods allowed for CORS. + * allow_headers: Headers allowed for CORS. + * ssl_certfile: SSL certificate file for HTTPS. See https://www.uvicorn.org/settings/#https. + * ssl_keyfile: SSL key file for HTTPS. See https://www.uvicorn.org/settings/#https. + * log_tokenization: Enable logging of parsed prompt tokens. + * patchmatch: Enable patchmatch inpaint code. + * models_dir: Path to the models directory. + * convert_cache_dir: Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). + * download_cache_dir: Path to the directory that contains dynamically downloaded models. + * legacy_conf_dir: Path to directory of legacy checkpoint config files. + * db_dir: Path to InvokeAI databases directory. + * outputs_dir: Path to directory for outputs. + * custom_nodes_dir: Path to directory for custom nodes. + * style_presets_dir: Path to directory for style presets. + * workflow_thumbnails_dir: Path to directory for workflow thumbnails. + * log_handlers: Log handler. Valid options are "console", "file=", "syslog=path|address:host:port", "http=". + * log_format: Log format. Use "plain" for text-only, "color" for colorized output, "legacy" for 2.3-style logging and "syslog" for syslog-style.
Valid values: `plain`, `color`, `syslog`, `legacy` + * log_level: Emit logging messages at this level or higher.
Valid values: `debug`, `info`, `warning`, `error`, `critical` + * log_sql: Log SQL queries. `log_level` must be `debug` for this to do anything. Extremely verbose. + * log_level_network: Log level for network-related messages. 'info' and 'debug' are very verbose.
Valid values: `debug`, `info`, `warning`, `error`, `critical` + * use_memory_db: Use in-memory database. Useful for development. + * dev_reload: Automatically reload when Python sources are changed. Does not reload node definitions. + * profile_graphs: Enable graph profiling using `cProfile`. + * profile_prefix: An optional prefix for profile output files. + * profiles_dir: Path to profiles output directory. + * max_cache_ram_gb: The maximum amount of CPU RAM to use for model caching in GB. If unset, the limit will be configured based on the available RAM. In most cases, it is recommended to leave this unset. + * max_cache_vram_gb: The amount of VRAM to use for model caching in GB. If unset, the limit will be configured based on the available VRAM and the device_working_mem_gb. In most cases, it is recommended to leave this unset. + * log_memory_usage: If True, a memory snapshot will be captured before and after every model cache operation, and the result will be logged (at debug level). There is a time cost to capturing the memory snapshots, so it is recommended to only enable this feature if you are actively inspecting the model cache's behaviour. + * model_cache_keep_alive_min: How long to keep models in cache after last use, in minutes. A value of 0 (the default) means models are kept in cache indefinitely. If no model generations occur within the timeout period, the model cache is cleared using the same logic as the 'Clear Model Cache' button. + * device_working_mem_gb: The amount of working memory to keep available on the compute device (in GB). Has no effect if running on CPU. If you are experiencing OOM errors, try increasing this value. + * enable_partial_loading: Enable partial loading of models. This enables models to run with reduced VRAM requirements (at the cost of slower speed) by streaming the model from RAM to VRAM as its used. In some edge cases, partial loading can cause models to run more slowly if they were previously being fully loaded into VRAM. + * keep_ram_copy_of_weights: Whether to keep a full RAM copy of a model's weights when the model is loaded in VRAM. Keeping a RAM copy increases average RAM usage, but speeds up model switching and LoRA patching (assuming there is sufficient RAM). Set this to False if RAM pressure is consistently high. + * ram: DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_ram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. + * vram: DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_vram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. + * lazy_offload: DEPRECATED: This setting is no longer used. Lazy-offloading is enabled by default. This config setting will be removed once the new model cache behavior is stable. + * pytorch_cuda_alloc_conf: Configure the Torch CUDA memory allocator. This will impact peak reserved VRAM usage and performance. Setting to "backend:cudaMallocAsync" works well on many systems. The optimal configuration is highly dependent on the system configuration (device type, VRAM, CUDA driver version, etc.), so must be tuned experimentally. + * device: Preferred execution device. `auto` will choose the device depending on the hardware platform and the installed torch capabilities.
Valid values: `auto`, `cpu`, `cuda`, `mps`, `cuda:N` (where N is a device number) + * precision: Floating point precision. `float16` will consume half the memory of `float32` but produce slightly lower-quality images. The `auto` setting will guess the proper precision based on your video card and operating system.
Valid values: `auto`, `float16`, `bfloat16`, `float32` + * sequential_guidance: Whether to calculate guidance in serial instead of in parallel, lowering memory requirements. + * attention_type: Attention type.
Valid values: `auto`, `normal`, `xformers`, `sliced`, `torch-sdp` + * attention_slice_size: Slice size, valid when attention_type=="sliced".
Valid values: `auto`, `balanced`, `max`, `1`, `2`, `3`, `4`, `5`, `6`, `7`, `8` + * force_tiled_decode: Whether to enable tiled VAE decode (reduces memory consumption with some performance penalty). + * pil_compress_level: The compress_level setting of PIL.Image.save(), used for PNG encoding. All settings are lossless. 0 = no compression, 1 = fastest with slightly larger filesize, 9 = slowest with smallest filesize. 1 is typically the best setting. + * max_queue_size: Maximum number of items in the session queue. + * clear_queue_on_startup: Empties session queue on startup. + * allow_nodes: List of nodes to allow. Omit to allow all. + * deny_nodes: List of nodes to deny. Omit to deny none. + * node_cache_size: How many cached nodes to keep in memory. + * hashing_algorithm: Model hashing algorthim for model installs. 'blake3_multi' is best for SSDs. 'blake3_single' is best for spinning disk HDDs. 'random' disables hashing, instead assigning a UUID to models. Useful when using a memory db to reduce model installation time, or if you don't care about storing stable hashes for models. Alternatively, any other hashlib algorithm is accepted, though these are not nearly as performant as blake3.
Valid values: `blake3_multi`, `blake3_single`, `random`, `md5`, `sha1`, `sha224`, `sha256`, `sha384`, `sha512`, `blake2b`, `blake2s`, `sha3_224`, `sha3_256`, `sha3_384`, `sha3_512`, `shake_128`, `shake_256` + * remote_api_tokens: List of regular expression and token pairs used when downloading models from URLs. The download URL is tested against the regex, and if it matches, the token is provided in as a Bearer token. + * scan_models_on_startup: Scan the models directory on startup, registering orphaned models. This is typically only used in conjunction with `use_memory_db` for testing purposes. + * unsafe_disable_picklescan: UNSAFE. Disable the picklescan security check during model installation. Recommended only for development and testing purposes. This will allow arbitrary code execution during model installation, so should never be used in production. + * allow_unknown_models: Allow installation of models that we are unable to identify. If enabled, models will be marked as `unknown` in the database, and will not have any metadata associated with them. If disabled, unknown models will be rejected during installation. + * multiuser: Enable multiuser support. When disabled, the application runs in single-user mode using a default system account with administrator privileges. When enabled, requires user authentication and authorization. + * strict_password_checking: Enforce strict password requirements. When True, passwords must contain uppercase, lowercase, and numbers. When False (default), any password is accepted but its strength (weak/moderate/strong) is reported to the user. + */ + InvokeAIAppConfig: { + /** + * Schema Version + * @description Schema version of the config file. This is not a user-configurable setting. + * @default 4.0.2 + */ + schema_version?: string; + /** + * Legacy Models Yaml Path + * @description Path to the legacy models.yaml file. This is not a user-configurable setting. + */ + legacy_models_yaml_path?: string | null; + /** + * Host + * @description IP address to bind to. Use `0.0.0.0` to serve to your local network. + * @default 127.0.0.1 + */ + host?: string; + /** + * Port + * @description Port to bind to. + * @default 9090 + */ + port?: number; + /** + * Allow Origins + * @description Allowed CORS origins. + * @default [] + */ + allow_origins?: string[]; + /** + * Allow Credentials + * @description Allow CORS credentials. + * @default true + */ + allow_credentials?: boolean; + /** + * Allow Methods + * @description Methods allowed for CORS. + * @default [ + * "*" + * ] + */ + allow_methods?: string[]; + /** + * Allow Headers + * @description Headers allowed for CORS. + * @default [ + * "*" + * ] + */ + allow_headers?: string[]; + /** + * Ssl Certfile + * @description SSL certificate file for HTTPS. See https://www.uvicorn.org/settings/#https. + */ + ssl_certfile?: string | null; + /** + * Ssl Keyfile + * @description SSL key file for HTTPS. See https://www.uvicorn.org/settings/#https. + */ + ssl_keyfile?: string | null; + /** + * Log Tokenization + * @description Enable logging of parsed prompt tokens. + * @default false + */ + log_tokenization?: boolean; + /** + * Patchmatch + * @description Enable patchmatch inpaint code. + * @default true + */ + patchmatch?: boolean; + /** + * Models Dir + * Format: path + * @description Path to the models directory. + * @default models + */ + models_dir?: string; + /** + * Convert Cache Dir + * Format: path + * @description Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). + * @default models\.convert_cache + */ + convert_cache_dir?: string; + /** + * Download Cache Dir + * Format: path + * @description Path to the directory that contains dynamically downloaded models. + * @default models\.download_cache + */ + download_cache_dir?: string; + /** + * Legacy Conf Dir + * Format: path + * @description Path to directory of legacy checkpoint config files. + * @default configs + */ + legacy_conf_dir?: string; + /** + * Db Dir + * Format: path + * @description Path to InvokeAI databases directory. + * @default databases + */ + db_dir?: string; + /** + * Outputs Dir + * Format: path + * @description Path to directory for outputs. + * @default outputs + */ + outputs_dir?: string; + /** + * Custom Nodes Dir + * Format: path + * @description Path to directory for custom nodes. + * @default nodes + */ + custom_nodes_dir?: string; + /** + * Style Presets Dir + * Format: path + * @description Path to directory for style presets. + * @default style_presets + */ + style_presets_dir?: string; + /** + * Workflow Thumbnails Dir + * Format: path + * @description Path to directory for workflow thumbnails. + * @default workflow_thumbnails + */ + workflow_thumbnails_dir?: string; + /** + * Log Handlers + * @description Log handler. Valid options are "console", "file=", "syslog=path|address:host:port", "http=". + * @default [ + * "console" + * ] + */ + log_handlers?: string[]; + /** + * Log Format + * @description Log format. Use "plain" for text-only, "color" for colorized output, "legacy" for 2.3-style logging and "syslog" for syslog-style. + * @default color + * @enum {string} + */ + log_format?: "plain" | "color" | "syslog" | "legacy"; + /** + * Log Level + * @description Emit logging messages at this level or higher. + * @default info + * @enum {string} + */ + log_level?: "debug" | "info" | "warning" | "error" | "critical"; + /** + * Log Sql + * @description Log SQL queries. `log_level` must be `debug` for this to do anything. Extremely verbose. + * @default false + */ + log_sql?: boolean; + /** + * Log Level Network + * @description Log level for network-related messages. 'info' and 'debug' are very verbose. + * @default warning + * @enum {string} + */ + log_level_network?: "debug" | "info" | "warning" | "error" | "critical"; + /** + * Use Memory Db + * @description Use in-memory database. Useful for development. + * @default false + */ + use_memory_db?: boolean; + /** + * Dev Reload + * @description Automatically reload when Python sources are changed. Does not reload node definitions. + * @default false + */ + dev_reload?: boolean; + /** + * Profile Graphs + * @description Enable graph profiling using `cProfile`. + * @default false + */ + profile_graphs?: boolean; + /** + * Profile Prefix + * @description An optional prefix for profile output files. + */ + profile_prefix?: string | null; + /** + * Profiles Dir + * Format: path + * @description Path to profiles output directory. + * @default profiles + */ + profiles_dir?: string; + /** + * Max Cache Ram Gb + * @description The maximum amount of CPU RAM to use for model caching in GB. If unset, the limit will be configured based on the available RAM. In most cases, it is recommended to leave this unset. + */ + max_cache_ram_gb?: number | null; + /** + * Max Cache Vram Gb + * @description The amount of VRAM to use for model caching in GB. If unset, the limit will be configured based on the available VRAM and the device_working_mem_gb. In most cases, it is recommended to leave this unset. + */ + max_cache_vram_gb?: number | null; + /** + * Log Memory Usage + * @description If True, a memory snapshot will be captured before and after every model cache operation, and the result will be logged (at debug level). There is a time cost to capturing the memory snapshots, so it is recommended to only enable this feature if you are actively inspecting the model cache's behaviour. + * @default false + */ + log_memory_usage?: boolean; + /** + * Model Cache Keep Alive Min + * @description How long to keep models in cache after last use, in minutes. A value of 0 (the default) means models are kept in cache indefinitely. If no model generations occur within the timeout period, the model cache is cleared using the same logic as the 'Clear Model Cache' button. + * @default 0 + */ + model_cache_keep_alive_min?: number; + /** + * Device Working Mem Gb + * @description The amount of working memory to keep available on the compute device (in GB). Has no effect if running on CPU. If you are experiencing OOM errors, try increasing this value. + * @default 3 + */ + device_working_mem_gb?: number; + /** + * Enable Partial Loading + * @description Enable partial loading of models. This enables models to run with reduced VRAM requirements (at the cost of slower speed) by streaming the model from RAM to VRAM as its used. In some edge cases, partial loading can cause models to run more slowly if they were previously being fully loaded into VRAM. + * @default false + */ + enable_partial_loading?: boolean; + /** + * Keep Ram Copy Of Weights + * @description Whether to keep a full RAM copy of a model's weights when the model is loaded in VRAM. Keeping a RAM copy increases average RAM usage, but speeds up model switching and LoRA patching (assuming there is sufficient RAM). Set this to False if RAM pressure is consistently high. + * @default true + */ + keep_ram_copy_of_weights?: boolean; + /** + * Ram + * @description DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_ram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. + */ + ram?: number | null; + /** + * Vram + * @description DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_vram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. + */ + vram?: number | null; + /** + * Lazy Offload + * @description DEPRECATED: This setting is no longer used. Lazy-offloading is enabled by default. This config setting will be removed once the new model cache behavior is stable. + * @default true + */ + lazy_offload?: boolean; + /** + * Pytorch Cuda Alloc Conf + * @description Configure the Torch CUDA memory allocator. This will impact peak reserved VRAM usage and performance. Setting to "backend:cudaMallocAsync" works well on many systems. The optimal configuration is highly dependent on the system configuration (device type, VRAM, CUDA driver version, etc.), so must be tuned experimentally. + */ + pytorch_cuda_alloc_conf?: string | null; + /** + * Device + * @description Preferred execution device. `auto` will choose the device depending on the hardware platform and the installed torch capabilities.
Valid values: `auto`, `cpu`, `cuda`, `mps`, `cuda:N` (where N is a device number) + * @default auto + */ + device?: string; + /** + * Precision + * @description Floating point precision. `float16` will consume half the memory of `float32` but produce slightly lower-quality images. The `auto` setting will guess the proper precision based on your video card and operating system. + * @default auto + * @enum {string} + */ + precision?: "auto" | "float16" | "bfloat16" | "float32"; + /** + * Sequential Guidance + * @description Whether to calculate guidance in serial instead of in parallel, lowering memory requirements. + * @default false + */ + sequential_guidance?: boolean; + /** + * Attention Type + * @description Attention type. + * @default auto + * @enum {string} + */ + attention_type?: "auto" | "normal" | "xformers" | "sliced" | "torch-sdp"; + /** + * Attention Slice Size + * @description Slice size, valid when attention_type=="sliced". + * @default auto + * @enum {unknown} + */ + attention_slice_size?: "auto" | "balanced" | "max" | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8; + /** + * Force Tiled Decode + * @description Whether to enable tiled VAE decode (reduces memory consumption with some performance penalty). + * @default false + */ + force_tiled_decode?: boolean; + /** + * Pil Compress Level + * @description The compress_level setting of PIL.Image.save(), used for PNG encoding. All settings are lossless. 0 = no compression, 1 = fastest with slightly larger filesize, 9 = slowest with smallest filesize. 1 is typically the best setting. + * @default 1 + */ + pil_compress_level?: number; + /** + * Max Queue Size + * @description Maximum number of items in the session queue. + * @default 10000 + */ + max_queue_size?: number; + /** + * Clear Queue On Startup + * @description Empties session queue on startup. + * @default false + */ + clear_queue_on_startup?: boolean; + /** + * Allow Nodes + * @description List of nodes to allow. Omit to allow all. + */ + allow_nodes?: string[] | null; + /** + * Deny Nodes + * @description List of nodes to deny. Omit to deny none. + */ + deny_nodes?: string[] | null; + /** + * Node Cache Size + * @description How many cached nodes to keep in memory. + * @default 512 + */ + node_cache_size?: number; + /** + * Hashing Algorithm + * @description Model hashing algorthim for model installs. 'blake3_multi' is best for SSDs. 'blake3_single' is best for spinning disk HDDs. 'random' disables hashing, instead assigning a UUID to models. Useful when using a memory db to reduce model installation time, or if you don't care about storing stable hashes for models. Alternatively, any other hashlib algorithm is accepted, though these are not nearly as performant as blake3. + * @default blake3_single + * @enum {string} + */ + hashing_algorithm?: "blake3_multi" | "blake3_single" | "random" | "md5" | "sha1" | "sha224" | "sha256" | "sha384" | "sha512" | "blake2b" | "blake2s" | "sha3_224" | "sha3_256" | "sha3_384" | "sha3_512" | "shake_128" | "shake_256"; + /** + * Remote Api Tokens + * @description List of regular expression and token pairs used when downloading models from URLs. The download URL is tested against the regex, and if it matches, the token is provided in as a Bearer token. + */ + remote_api_tokens?: components["schemas"]["URLRegexTokenPair"][] | null; + /** + * Scan Models On Startup + * @description Scan the models directory on startup, registering orphaned models. This is typically only used in conjunction with `use_memory_db` for testing purposes. + * @default false + */ + scan_models_on_startup?: boolean; + /** + * Unsafe Disable Picklescan + * @description UNSAFE. Disable the picklescan security check during model installation. Recommended only for development and testing purposes. This will allow arbitrary code execution during model installation, so should never be used in production. + * @default false + */ + unsafe_disable_picklescan?: boolean; + /** + * Allow Unknown Models + * @description Allow installation of models that we are unable to identify. If enabled, models will be marked as `unknown` in the database, and will not have any metadata associated with them. If disabled, unknown models will be rejected during installation. + * @default true + */ + allow_unknown_models?: boolean; + /** + * Multiuser + * @description Enable multiuser support. When disabled, the application runs in single-user mode using a default system account with administrator privileges. When enabled, requires user authentication and authorization. + * @default false + */ + multiuser?: boolean; + /** + * Strict Password Checking + * @description Enforce strict password requirements. When True, passwords must contain uppercase, lowercase, and numbers. When False (default), any password is accepted but its strength (weak/moderate/strong) is reported to the user. + * @default false + */ + strict_password_checking?: boolean; + }; + /** + * InvokeAIAppConfigWithSetFields + * @description InvokeAI App Config with model fields set + */ + InvokeAIAppConfigWithSetFields: { + /** + * Set Fields + * @description The set fields + */ + set_fields: string[]; + /** @description The InvokeAI App Config */ + config: components["schemas"]["InvokeAIAppConfig"]; + }; + /** + * Adjust Image Hue Plus + * @description Adjusts the Hue of an image by rotating it in the selected color space. Originally created by @dwringer + */ + InvokeAdjustImageHuePlusInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to adjust + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Space + * @description Color space in which to rotate hue by polar coords (*: non-invertible) + * @default HSV / HSL / RGB + * @enum {string} + */ + space?: "HSV / HSL / RGB" | "Okhsl" | "Okhsv" | "*Oklch / Oklab" | "*LCh / CIELab" | "*UPLab (w/CIELab_to_UPLab.icc)"; + /** + * Degrees + * @description Degrees by which to rotate image hue + * @default 0 + */ + degrees?: number; + /** + * Preserve Lightness + * @description Whether to preserve CIELAB lightness values + * @default false + */ + preserve_lightness?: boolean; + /** + * Ok Adaptive Gamut + * @description Higher preserves chroma at the expense of lightness (Oklab) + * @default 0.05 + */ + ok_adaptive_gamut?: number; + /** + * Ok High Precision + * @description Use more steps in computing gamut (Oklab/Okhsv/Okhsl) + * @default true + */ + ok_high_precision?: boolean; + /** + * type + * @default invokeai_img_hue_adjust_plus + * @constant + */ + type: "invokeai_img_hue_adjust_plus"; + }; + /** + * Equivalent Achromatic Lightness + * @description Calculate Equivalent Achromatic Lightness from image. Originally created by @dwringer + */ + InvokeEquivalentAchromaticLightnessInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Image from which to get channel + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default invokeai_ealightness + * @constant + */ + type: "invokeai_ealightness"; + }; + /** + * Image Layer Blend + * @description Blend two images together, with optional opacity, mask, and blend modes. Originally created by @dwringer + */ + InvokeImageBlendInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The top image to blend + * @default null + */ + layer_upper?: components["schemas"]["ImageField"] | null; + /** + * Blend Mode + * @description Available blend modes + * @default Normal + * @enum {string} + */ + blend_mode?: "Normal" | "Lighten Only" | "Darken Only" | "Lighten Only (EAL)" | "Darken Only (EAL)" | "Hue" | "Saturation" | "Color" | "Luminosity" | "Linear Dodge (Add)" | "Subtract" | "Multiply" | "Divide" | "Screen" | "Overlay" | "Linear Burn" | "Difference" | "Hard Light" | "Soft Light" | "Vivid Light" | "Linear Light" | "Color Burn" | "Color Dodge"; + /** + * Opacity + * @description Desired opacity of the upper layer + * @default 1 + */ + opacity?: number; + /** + * @description Optional mask, used to restrict areas from blending + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * Fit To Width + * @description Scale upper layer to fit base width + * @default false + */ + fit_to_width?: boolean; + /** + * Fit To Height + * @description Scale upper layer to fit base height + * @default true + */ + fit_to_height?: boolean; + /** + * @description The bottom image to blend + * @default null + */ + layer_base?: components["schemas"]["ImageField"] | null; + /** + * Color Space + * @description Available color spaces for blend computations + * @default RGB + * @enum {string} + */ + color_space?: "RGB" | "Linear RGB" | "HSL (RGB)" | "HSV (RGB)" | "Okhsl" | "Okhsv" | "Oklch (Oklab)" | "LCh (CIELab)"; + /** + * Adaptive Gamut + * @description Adaptive gamut clipping (0=off). Higher prioritizes chroma over lightness + * @default 0 + */ + adaptive_gamut?: number; + /** + * High Precision + * @description Use more steps in computing gamut when possible + * @default true + */ + high_precision?: boolean; + /** + * type + * @default invokeai_img_blend + * @constant + */ + type: "invokeai_img_blend"; + }; + /** + * Image Compositor + * @description Removes backdrop from subject image then overlays subject on background image. Originally created by @dwringer + */ + InvokeImageCompositorInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Image of the subject on a plain monochrome background + * @default null + */ + image_subject?: components["schemas"]["ImageField"] | null; + /** + * @description Image of a background scene + * @default null + */ + image_background?: components["schemas"]["ImageField"] | null; + /** + * Chroma Key + * @description Can be empty for corner flood select, or CSS-3 color or tuple + * @default + */ + chroma_key?: string; + /** + * Threshold + * @description Subject isolation flood-fill threshold + * @default 50 + */ + threshold?: number; + /** + * Fill X + * @description Scale base subject image to fit background width + * @default false + */ + fill_x?: boolean; + /** + * Fill Y + * @description Scale base subject image to fit background height + * @default true + */ + fill_y?: boolean; + /** + * X Offset + * @description x-offset for the subject + * @default 0 + */ + x_offset?: number; + /** + * Y Offset + * @description y-offset for the subject + * @default 0 + */ + y_offset?: number; + /** + * type + * @default invokeai_img_composite + * @constant + */ + type: "invokeai_img_composite"; + }; + /** + * Image Dilate or Erode + * @description Dilate (expand) or erode (contract) an image. Originally created by @dwringer + */ + InvokeImageDilateOrErodeInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image from which to create a mask + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Lightness Only + * @description If true, only applies to image lightness (CIELa*b*) + * @default false + */ + lightness_only?: boolean; + /** + * Radius W + * @description Width (in pixels) by which to dilate(expand) or erode (contract) the image + * @default 4 + */ + radius_w?: number; + /** + * Radius H + * @description Height (in pixels) by which to dilate(expand) or erode (contract) the image + * @default 4 + */ + radius_h?: number; + /** + * Mode + * @description How to operate on the image + * @default Dilate + * @enum {string} + */ + mode?: "Dilate" | "Erode"; + /** + * type + * @default invokeai_img_dilate_erode + * @constant + */ + type: "invokeai_img_dilate_erode"; + }; + /** + * Enhance Image + * @description Applies processing from PIL's ImageEnhance module. Originally created by @dwringer + */ + InvokeImageEnhanceInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image for which to apply processing + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Invert + * @description Whether to invert the image colors + * @default false + */ + invert?: boolean; + /** + * Color + * @description Color enhancement factor + * @default 1 + */ + color?: number; + /** + * Contrast + * @description Contrast enhancement factor + * @default 1 + */ + contrast?: number; + /** + * Brightness + * @description Brightness enhancement factor + * @default 1 + */ + brightness?: number; + /** + * Sharpness + * @description Sharpness enhancement factor + * @default 1 + */ + sharpness?: number; + /** + * type + * @default invokeai_img_enhance + * @constant + */ + type: "invokeai_img_enhance"; + }; + /** + * Image Value Thresholds + * @description Clip image to pure black/white past specified thresholds. Originally created by @dwringer + */ + InvokeImageValueThresholdsInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image from which to create a mask + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Invert Output + * @description Make light areas dark and vice versa + * @default false + */ + invert_output?: boolean; + /** + * Renormalize Values + * @description Rescale remaining values from minimum to maximum + * @default false + */ + renormalize_values?: boolean; + /** + * Lightness Only + * @description If true, only applies to image lightness (CIELa*b*) + * @default false + */ + lightness_only?: boolean; + /** + * Threshold Upper + * @description Threshold above which will be set to full value + * @default 0.5 + */ + threshold_upper?: number; + /** + * Threshold Lower + * @description Threshold below which will be set to minimum value + * @default 0.5 + */ + threshold_lower?: number; + /** + * type + * @default invokeai_img_val_thresholds + * @constant + */ + type: "invokeai_img_val_thresholds"; + }; + /** + * ItemIdsResult + * @description Response containing ordered item ids with metadata for optimistic updates. + */ + ItemIdsResult: { + /** + * Item Ids + * @description Ordered list of item ids + */ + item_ids: number[]; + /** + * Total Count + * @description Total number of queue items matching the query + */ + total_count: number; + }; + /** + * IterateInvocation + * @description Iterates over a list of items + */ + IterateInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Collection + * @description The list of items to iterate over + * @default [] + */ + collection?: unknown[]; + /** + * Index + * @description The index, will be provided on executed iterators + * @default 0 + */ + index?: number; + /** + * type + * @default iterate + * @constant + */ + type: "iterate"; + }; + /** + * IterateInvocationOutput + * @description Used to connect iteration outputs. Will be expanded to a specific output. + */ + IterateInvocationOutput: { + /** + * Collection Item + * @description The item being iterated over + */ + item: unknown; + /** + * Index + * @description The index of the item + */ + index: number; + /** + * Total + * @description The total number of items + */ + total: number; + /** + * type + * @default iterate_output + * @constant + */ + type: "iterate_output"; + }; + /** + * JsonListStringsOutput + * @description Output class for two strings extracted from a JSON list. + */ + JsonListStringsOutput: { + /** + * Prompt + * @description The prompt string from the JSON list + */ + prompt: string; + /** + * Seed Vector + * @description The seed vector string from the JSON list + */ + seed_vector: string; + /** + * type + * @default json_list_strings_output + * @constant + */ + type: "json_list_strings_output"; + }; + JsonValue: unknown; + /** + * LaMa Infill + * @description Infills transparent areas of an image using the LaMa model + */ + LaMaInfillInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to process + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default infill_lama + * @constant + */ + type: "infill_lama"; + }; + /** + * Latent Average + * @description Average two latents + */ + LatentAverageInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latentA?: components["schemas"]["LatentsField"] | null; + /** + * @description Latents tensor + * @default null + */ + latentB?: components["schemas"]["LatentsField"] | null; + /** + * Channels + * @description Comma-separated list of channel indices to average. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * type + * @default latent_average + * @constant + */ + type: "latent_average"; + }; + /** + * Latent Band Pass Filter + * @description Latent Band Pass Filter + */ + LatentBandPassFilterInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Input latent + * @default null + */ + input_latent?: components["schemas"]["LatentsField"] | null; + /** + * Lower Freq + * @description lower frequency (0-0.5) + * @default 0.1 + */ + lower_freq?: number; + /** + * Upper Freq + * @description upper frequency (0-0.5) + * @default 0.2 + */ + upper_freq?: number; + /** + * Normalize Std Dev + * @description normalize std dev to 1.0 + * @default false + */ + normalize_std_dev?: boolean; + /** + * type + * @default latent_band_pass_filter + * @constant + */ + type: "latent_band_pass_filter"; + }; + /** + * Latent Blend Linear + * @description Blend two latents Linearly + */ + LatentBlendLinearInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latentA?: components["schemas"]["LatentsField"] | null; + /** + * @description Latents tensor + * @default null + */ + latentB?: components["schemas"]["LatentsField"] | null; + /** + * A B Ratio + * @description Ratio of latentA to latentB (1=All A, 0=All B) + * @default 0.5 + */ + a_b_ratio?: number; + /** + * Channels + * @description Comma-separated list of channel indices to average. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * type + * @default latent_blend_linear + * @constant + */ + type: "latent_blend_linear"; + }; + /** + * Latent Channels to Grid + * @description Generates a scaled grid Images from latent channels + */ + LatentChannelsToGridInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Disable + * @description Disable this node + * @default false + */ + disable?: boolean; + /** + * @description Latents tensor + * @default null + */ + latent?: components["schemas"]["LatentsField"] | null; + /** + * Scale + * @description Overall scale factor for the grid. + * @default 1 + */ + scale?: number; + /** + * Normalize Channels + * @description Normalize all channels using a common min/max range. + * @default false + */ + normalize_channels?: boolean; + /** + * Output Bit Depth + * @description Output as 8-bit or 16-bit grayscale. + * @default 8 + * @enum {string} + */ + output_bit_depth?: "8" | "16"; + /** + * Title + * @description Title to display on the image + * @default Latent Channel Grid + */ + title?: string; + /** + * type + * @default latent_channels_to_grid + * @constant + */ + type: "latent_channels_to_grid"; + }; + /** + * Latent Combine + * @description Combines two latent tensors using various methods + */ + LatentCombineInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latentA?: components["schemas"]["LatentsField"] | null; + /** + * @description Latents tensor + * @default null + */ + latentB?: components["schemas"]["LatentsField"] | null; + /** + * Method + * @description Combination method + * @default average + * @enum {string} + */ + method?: "average" | "add" | "subtract" | "difference" | "maximum" | "minimum" | "multiply" | "frequency_blend" | "screen" | "dodge" | "burn" | "overlay" | "soft_light" | "hard_light" | "color_dodge" | "color_burn" | "linear_dodge" | "linear_burn"; + /** + * Weighted Combine + * @description Alter input latents by the provided weights + * @default false + */ + weighted_combine?: boolean; + /** + * Weight A + * @description Weight for latent A + * @default 0.5 + */ + weight_a?: number; + /** + * Weight B + * @description Weight for latent B + * @default 0.5 + */ + weight_b?: number; + /** + * Scale To Input Ranges + * @description Scale output to input ranges + * @default false + */ + scale_to_input_ranges?: boolean; + /** + * Channels + * @description Comma-separated list of channel indices to combine. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * type + * @default latent_combine + * @constant + */ + type: "latent_combine"; + }; + /** + * Latent Dtype Convert + * @description Converts the dtype of a latent tensor. + */ + LatentDtypeConvertInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latent?: components["schemas"]["LatentsField"] | null; + /** + * Dtype + * @description The target dtype for the latent tensor. + * @default bfloat16 + * @enum {string} + */ + dtype?: "float32" | "float16" | "bfloat16"; + /** + * type + * @default latent_dtype_convert + * @constant + */ + type: "latent_dtype_convert"; + }; + /** + * Latent High Pass Filter + * @description Latent High Pass Filter + */ + LatentHighPassFilterInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Input latent + * @default null + */ + input_latent?: components["schemas"]["LatentsField"] | null; + /** + * Lower Freq + * @description lower frequency (0-0.5) + * @default 0.15 + */ + lower_freq?: number; + /** + * Normalize Std Dev + * @description normalize std dev to 1.0 + * @default false + */ + normalize_std_dev?: boolean; + /** + * type + * @default latent_high_pass_filter + * @constant + */ + type: "latent_high_pass_filter"; + }; + /** + * Latent Low Pass Filter + * @description Latent Low Pass Filter + */ + LatentLowPassFilterInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Input latent + * @default null + */ + input_latent?: components["schemas"]["LatentsField"] | null; + /** + * Upper Freq + * @description upper frequency (0-0.5) + * @default 0.05 + */ + upper_freq?: number; + /** + * Normalize Std Dev + * @description normalize std dev to 1.0 + * @default false + */ + normalize_std_dev?: boolean; + /** + * type + * @default latent_low_pass_filter + * @constant + */ + type: "latent_low_pass_filter"; + }; + /** + * Latent Match + * @description Matches the distribution of one latent tensor to that of another, + * using either histogram matching or standard deviation matching. + */ + LatentMatchInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The input latent tensor to be matched. + * @default null + */ + input_latent?: components["schemas"]["LatentsField"] | null; + /** + * @description The reference latent tensor to match against. + * @default null + */ + reference_latent?: components["schemas"]["LatentsField"] | null; + /** + * Method + * @description The matching method to use. + * @default histogram + * @enum {string} + */ + method?: "histogram" | "std_dev" | "mean" | "std_dev+mean" | "cdf" | "moment" | "range" | "std_dev+mean+range"; + /** + * Channels + * @description Comma-separated list of channel indices to match. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * type + * @default latent_match + * @constant + */ + type: "latent_match"; + }; + /** + * Latent Modify + * @description Modifies selected channels of a latent tensor using scale and shift. + */ + LatentModifyChannelsInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latent?: components["schemas"]["LatentsField"] | null; + /** + * Channels + * @description Comma-separated list of channel indices to modify. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * Scale + * @description Scale factor to apply to the selected channels. + * @default 1 + */ + scale?: number; + /** + * Shift + * @description Shift value to add to the selected channels. + * @default 0 + */ + shift?: number; + /** + * type + * @default latent_modify_channels + * @constant + */ + type: "latent_modify_channels"; + }; + /** + * Latent Normalize Range + * @description Normalize a Latents Range + */ + LatentNormalizeRangeInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latent?: components["schemas"]["LatentsField"] | null; + /** + * Min + * @description Min for new range + * @default -1 + */ + min?: number; + /** + * Max + * @description Max for new range + * @default 1 + */ + max?: number; + /** + * Channels + * @description Comma-separated list of channel indices to average. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * type + * @default latent_normalize_range + * @constant + */ + type: "latent_normalize_range"; + }; + /** + * Latent Normalize Std-Dev + * @description Normalize a Latents Std Dev + */ + LatentNormalizeStdDevInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latent?: components["schemas"]["LatentsField"] | null; + /** + * Std Dev + * @description Standard deviation to normalize to + * @default 1 + */ + std_dev?: number; + /** + * Channels + * @description Comma-separated list of channel indices to average. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * type + * @default latent_normalize_std_dev + * @constant + */ + type: "latent_normalize_std_dev"; + }; + /** + * Latent Normalize STD Dev and Range + * @description Normalize a Latents Range and Std Dev + */ + LatentNormalizeStdRangeInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latent?: components["schemas"]["LatentsField"] | null; + /** + * Std Dev + * @description Standard deviation to normalize to + * @default 1 + */ + std_dev?: number; + /** + * Min + * @description Min for new range + * @default -1 + */ + min?: number; + /** + * Max + * @description Max for new range + * @default 1 + */ + max?: number; + /** + * Tolerance + * @description tolerance + * @default 0.001 + */ + tolerance?: number; + /** + * Max Iterations + * @description Max iterations + * @default 10 + */ + max_iterations?: number; + /** + * Channels + * @description Comma-separated list of channel indices to average. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * type + * @default latent_normalize_std_dev_range + * @constant + */ + type: "latent_normalize_std_dev_range"; + }; + /** + * Latent Plot + * @description Generates plots from latent channels and display in a grid. + */ + LatentPlotInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Disable + * @description Disable this node + * @default false + */ + disable?: boolean; + /** + * @description Latents tensor + * @default null + */ + latent?: components["schemas"]["LatentsField"] | null; + /** + * Cell X Multiplier + * @description x size multiplier for grid cells + * @default 4 + */ + cell_x_multiplier?: number; + /** + * Cell Y Multiplier + * @description y size multiplier for grid cells + * @default 3 + */ + cell_y_multiplier?: number; + /** + * Histogram Plot + * @description Plot histogram + * @default true + */ + histogram_plot?: boolean; + /** + * Histogram Bins + * @description Number of bins for the histogram + * @default 256 + */ + histogram_bins?: number; + /** + * Histogram Log Scale + * @description Use log scale for histogram y-axis + * @default false + */ + histogram_log_scale?: boolean; + /** + * Box Plot + * @description Plot box and whisker + * @default true + */ + box_plot?: boolean; + /** + * Stats Plot + * @description Plot distribution data (mean, std, mode, min, max, dtype) + * @default true + */ + stats_plot?: boolean; + /** + * Title + * @description Title to display on the image + * @default Latent Channel Plots + */ + title?: string; + /** + * Channels + * @description Comma-separated list of channel indices to plot. Use 'all' for all channels. + * @default all + */ + channels?: string; + /** + * Common Axis + * @description Use a common axis scales for all plots + * @default false + */ + common_axis?: boolean; + /** + * type + * @default latent_plot + * @constant + */ + type: "latent_plot"; + }; + /** + * Latent Quantize (Kohonen map) + * @description Use a self-organizing map to quantize the values of a latent tensor. + * + * This is highly experimental and not really suitable for most use cases. It's very easy to use settings that will appear to hang, or tie up the PC for a very long time, so use of this node is somewhat discouraged. + */ + LatentSOMInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The latents tensor to quantize + * @default null + */ + latents_in?: components["schemas"]["LatentsField"] | null; + /** + * @description Optional alternate latents to use for training + * @default null + */ + reference_in?: components["schemas"]["LatentsField"] | null; + /** + * Width + * @description Width (in cells) of the self-organizing map + * @default 4 + */ + width?: number; + /** + * Height + * @description Height (in cells) of the self-organizing map + * @default 3 + */ + height?: number; + /** + * Steps + * @description Training step count for the self-organizing map + * @default 256 + */ + steps?: number; + /** + * type + * @default latent_som + * @constant + */ + type: "latent_som"; + }; + /** + * Latent White Noise + * @description Creates White Noise latent. + */ + LatentWhiteNoiseInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Width + * @description Desired image width + * @default 512 + */ + width?: number; + /** + * Height + * @description Desired image height + * @default 512 + */ + height?: number; + /** + * Latent Type + * @description Latent Type + * @default Flux + * @enum {string} + */ + latent_type?: "Flux" | "SD3.5" | "SD/SDXL"; + /** + * Noise Type + * @description Noise Type (Uniform=random, Gausian=Normal distribution) + * @default Gausian + * @enum {string} + */ + noise_type?: "Uniform" | "Gausian"; + /** + * Seed + * @description Seed for noise generation + * @default 0 + */ + seed?: number; + /** + * type + * @default latent_white_noise + * @constant + */ + type: "latent_white_noise"; + }; + /** + * Latents Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option + */ + LatentsCollectionIndexInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * Random + * @description Random Index? + * @default true + */ + random?: boolean; + /** + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 + */ + index?: number; + /** + * Collection + * @description latents collection + * @default null + */ + collection?: components["schemas"]["LatentsField"][] | null; + /** + * type + * @default latents_collection_index + * @constant + */ + type: "latents_collection_index"; + }; + /** + * Latents Collection Primitive + * @description A collection of latents tensor primitive values + */ + LatentsCollectionInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Collection + * @description The collection of latents tensors + * @default null + */ + collection?: components["schemas"]["LatentsField"][] | null; + /** + * type + * @default latents_collection + * @constant + */ + type: "latents_collection"; + }; + /** + * Latents Collection Primitive Linked + * @description A collection of latents tensor primitive values + */ + LatentsCollectionLinkedInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Collection + * @description The collection of latents tensors + * @default null + */ + collection?: components["schemas"]["LatentsField"][] | null; + /** + * type + * @default latents_collection_linked + * @constant + */ + type: "latents_collection_linked"; + /** + * @description The latents tensor + * @default null + */ + latents?: components["schemas"]["LatentsField"] | null; + }; + /** + * LatentsCollectionOutput + * @description Base class for nodes that output a collection of latents tensors + */ + LatentsCollectionOutput: { + /** + * Collection + * @description Latents tensor + */ + collection: components["schemas"]["LatentsField"][]; + /** + * type + * @default latents_collection_output + * @constant + */ + type: "latents_collection_output"; + }; + /** + * LatentsField + * @description A latents tensor primitive field + */ + LatentsField: { + /** + * Latents Name + * @description The name of the latents + */ + latents_name: string; + /** + * Seed + * @description Seed used to generate this latents + * @default null + */ + seed?: number | null; + }; + /** + * Latents Primitive + * @description A latents tensor primitive value + */ + LatentsInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The latents tensor + * @default null + */ + latents?: components["schemas"]["LatentsField"] | null; + /** + * type + * @default latents + * @constant + */ + type: "latents"; + }; + /** + * LatentsMetaOutput + * @description Latents + metadata + */ + LatentsMetaOutput: { + /** @description Metadata Dict */ + metadata: components["schemas"]["MetadataField"]; + /** + * type + * @default latents_meta_output + * @constant + */ + type: "latents_meta_output"; + /** @description Latents tensor */ + latents: components["schemas"]["LatentsField"]; + /** + * Width + * @description Width of output (px) + */ + width: number; + /** + * Height + * @description Height of output (px) + */ + height: number; + }; + /** + * LatentsOutput + * @description Base class for nodes that output a single latents tensor + */ + LatentsOutput: { + /** @description Latents tensor */ + latents: components["schemas"]["LatentsField"]; + /** + * Width + * @description Width of output (px) + */ + width: number; + /** + * Height + * @description Height of output (px) + */ + height: number; + /** + * type + * @default latents_output + * @constant + */ + type: "latents_output"; + }; + /** + * Latents to Image - SD1.5, SDXL + * @description Generates an image from latents. + */ + LatentsToImageInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latents?: components["schemas"]["LatentsField"] | null; + /** + * @description VAE + * @default null + */ + vae?: components["schemas"]["VAEField"] | null; + /** + * Tiled + * @description Processing using overlapping tiles (reduce memory consumption) + * @default false + */ + tiled?: boolean; + /** + * Tile Size + * @description The tile size for VAE tiling in pixels (image space). If set to 0, the default tile size for the model will be used. Larger tile sizes generally produce better results at the cost of higher memory usage. + * @default 0 + */ + tile_size?: number; + /** + * Fp32 + * @description Whether or not to use full float32 precision + * @default false + */ + fp32?: boolean; + /** + * type + * @default l2i + * @constant + */ + type: "l2i"; + }; + /** + * Lens Aperture Generator + * @description Generates a grayscale aperture shape with adjustable number of blades, curvature, rotation, and softness. + */ + LensApertureGeneratorInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Number Of Blades + * @description Number of aperture blades (e.g., 5 for pentagon, 6 for hexagon) + * @default 6 + */ + number_of_blades?: number; + /** + * Rounded Blades + * @description Use curved edges between aperture blades + * @default true + */ + rounded_blades?: boolean; + /** + * Curvature + * @description Aperture blade curvature (1.0 is circular) + * @default 0.5 + */ + curvature?: number; + /** + * Rotation + * @description Clockwise rotation of the aperture shape in degrees + * @default 0 + */ + rotation?: number; + /** + * Softness + * @description Softness of the aperture edge (0.0 = hard, 1.0 = feathered) + * @default 0.1 + */ + softness?: number; + /** + * type + * @default lens_aperture_generator + * @constant + */ + type: "lens_aperture_generator"; + }; + /** + * Lens Blur + * @description Adds lens blur to the input image, first converting the input image to RGB + */ + LensBlurInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to streak + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * @description The depth map to use + * @default null + */ + depth_map?: components["schemas"]["ImageField"] | null; + /** + * @description The aperture image to use for convolution + * @default null + */ + aperture_image?: components["schemas"]["ImageField"] | null; + /** + * Focal Distance + * @description The distance at which focus is sharp: 0.0 (near) - 1.0 (far) + * @default 0.5 + */ + focal_distance?: number; + /** + * Distance Range + * @description The depth of field range around the focal distance + * @default 0.2 + */ + distance_range?: number; + /** + * Max Blur Radius + * @description Maximum blur radius + * @default 5 + */ + max_blur_radius?: number; + /** + * Max Blur Steps + * @description Number of blur maps to use internally + * @default 32 + */ + max_blur_steps?: number; + /** + * Anamorphic Factor + * @description Anamorphic squeeze factor + * @default 1.33 + */ + anamorphic_factor?: number; + /** + * Highlight Threshold + * @description The luminance threshold at which highlight enhancement begins + * @default 0.75 + */ + highlight_threshold?: number; + /** + * Highlight Factor + * @description Minimum multiplier for highlights at the threshold + * @default 1 + */ + highlight_factor?: number; + /** + * Highlight Factor High + * @description Maximum multiplier for highlights at full brightness + * @default 2 + */ + highlight_factor_high?: number; + /** + * Reduce Halos + * @description Reduce halos around areas of sharp depth distance (SLOW) + * @default true + */ + reduce_halos?: boolean; + /** + * type + * @default lens_blur + * @constant + */ + type: "lens_blur"; + }; + /** + * Lens Vignette + * @description Apply realistic lens vignetting to an image with configurable parameters. + */ + LensVignetteInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to apply vignetting to + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Intensity + * @description Intensity of the vignette effect (0-4) + * @default 0.5 + */ + intensity?: number; + /** + * Aperture Factor + * @description Aperture falloff factor (>1.2 = strong, <0.5 = weak) + * @default 0.5 + */ + aperture_factor?: number; + /** + * Center X + * @description X coordinate of vignette center (0-1) + * @default 0.5 + */ + center_x?: number; + /** + * Center Y + * @description Y coordinate of vignette center (0-1) + * @default 0.5 + */ + center_y?: number; + /** + * Preserve Highlights + * @description Preserve highlights + * @default true + */ + preserve_highlights?: boolean; + /** + * type + * @default lens_vignette + * @constant + */ + type: "lens_vignette"; + }; + /** + * Lineart Anime Edge Detection + * @description Geneartes an edge map using the Lineart model. + */ + LineartAnimeEdgeDetectionInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to process + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default lineart_anime_edge_detection + * @constant + */ + type: "lineart_anime_edge_detection"; + }; + /** + * Lineart Edge Detection + * @description Generates an edge map using the Lineart model. + */ + LineartEdgeDetectionInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to process + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Coarse + * @description Whether to use coarse mode + * @default false + */ + coarse?: boolean; + /** + * type + * @default lineart_edge_detection + * @constant + */ + type: "lineart_edge_detection"; + }; + /** + * LLaVA OneVision VLLM + * @description Run a LLaVA OneVision VLLM model. + */ + LlavaOnevisionVllmInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Images + * @description Input image. + * @default null + */ + images?: (components["schemas"]["ImageField"][] | components["schemas"]["ImageField"]) | null; + /** + * Prompt + * @description Input text prompt. + * @default + */ + prompt?: string; + /** + * LLaVA Model Type + * @description The VLLM model to use + * @default null + */ + vllm_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default llava_onevision_vllm + * @constant + */ + type: "llava_onevision_vllm"; + }; + /** + * LlavaOnevision_Diffusers_Config + * @description Model config for Llava Onevision models. + */ + LlavaOnevision_Diffusers_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + /** + * Type + * @default llava_onevision + * @constant + */ + type: "llava_onevision"; + /** + * Base + * @default any + * @constant + */ + base: "any"; + /** + * Cpu Only + * @description Whether this model should run on CPU only + */ + cpu_only: boolean | null; + }; + /** + * LoRA Collection From Path + * @description Loads a collection of LoRAs filtered based on their path on disk. + */ + LoRACollectionFromPathInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRA Type + * @description The type of LoRAs that will be retrieved. + * @default sdxl + * @enum {string} + */ + model_type?: "sd-1" | "sd-2" | "sdxl" | "sdxl-refiner"; + /** + * Path Filter + * @description The path for the lora must contain the filter string + * @default null + */ + path_filter?: string | null; + /** + * Default Weight + * @description The weight to apply to each lora. This can be changed after using a Reapply LoRA Weight node. + * @default 0.75 + */ + default_weight?: number; + /** + * type + * @default lora_collection_from_path_invocation + * @constant + */ + type: "lora_collection_from_path_invocation"; + }; + /** + * LoRACollectionFromPathOutput + * @description Filtered LoRA Colleciton Output + */ + LoRACollectionFromPathOutput: { + /** + * LoRAs + * @description A collection of LoRAs. + */ + loras: components["schemas"]["LoRAField"][]; + /** + * type + * @default lora_collection_from_path_output + * @constant + */ + type: "lora_collection_from_path_output"; + }; + /** + * LoRA Collection Primitive + * @description A collection of LoRA primitive values + */ + LoRACollectionInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRAs + * @description The collection of LoRA values + * @default [] + */ + collection?: components["schemas"]["LoRAField"][]; + /** + * type + * @default lora_collection + * @constant + */ + type: "lora_collection"; + }; + /** + * LoRA Collection Primitive Linked + * @description Selects a LoRA model and weight. + */ + LoRACollectionLinkedInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRAs + * @description The collection of LoRA values + * @default [] + */ + collection?: components["schemas"]["LoRAField"][]; + /** + * type + * @default lora_collection_linked + * @constant + */ + type: "lora_collection_linked"; + /** + * LoRA + * @description LoRA model to load + * @default null + */ + lora?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 + */ + weight?: number; + }; + /** + * Apply LoRA Collection - SD1.5 + * @description Applies a collection of LoRAs to the provided UNet and CLIP models. + */ + LoRACollectionLoader: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. + * @default null + */ + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; + /** + * UNet + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet?: components["schemas"]["UNetField"] | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * type + * @default lora_collection_loader + * @constant + */ + type: "lora_collection_loader"; + }; + /** LoRACollectionOutput */ + LoRACollectionOutput: { + /** + * LoRAs + * @description The collection of input items + */ + collection: components["schemas"]["LoRAField"][]; + /** + * type + * @default lora_collection_output + * @constant + */ + type: "lora_collection_output"; + }; + /** + * LoRA Collection Toggle + * @description Allows boolean selection between two separate LoRA collection inputs + */ + LoRACollectionToggleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * Col1 + * @description First LoRA Collection Input + * @default null + */ + col1?: components["schemas"]["LoRAField"][] | null; + /** + * Col2 + * @description Second LoRA Collection Input + * @default null + */ + col2?: components["schemas"]["LoRAField"][] | null; + /** + * type + * @default lora_collection_toggle + * @constant + */ + type: "lora_collection_toggle"; + }; + /** LoRACollectionToggleOutput */ + LoRACollectionToggleOutput: { + /** + * LoRA Collection + * @description Collection of LoRA models and weights + */ + collection: components["schemas"]["LoRAField"][]; + /** + * type + * @default lora_collection_toggle_output + * @constant + */ + type: "lora_collection_toggle_output"; + }; + /** LoRAField */ + LoRAField: { + /** @description Info to load lora model */ + lora: components["schemas"]["ModelIdentifierField"]; + /** + * Weight + * @description Weight to apply to lora model + */ + weight: number; + }; + /** + * Apply LoRA - SD1.5 + * @description Apply selected lora to unet and text_encoder. + */ + LoRALoaderInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRA + * @description LoRA model to load + * @default null + */ + lora?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 + */ + weight?: number; + /** + * UNet + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet?: components["schemas"]["UNetField"] | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * type + * @default lora_loader + * @constant + */ + type: "lora_loader"; + }; + /** + * LoRALoaderOutput + * @description Model loader output + */ + LoRALoaderOutput: { + /** + * UNet + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet: components["schemas"]["UNetField"] | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip: components["schemas"]["CLIPField"] | null; + /** + * type + * @default lora_loader_output + * @constant + */ + type: "lora_loader_output"; + }; + /** + * LoRAMetadataField + * @description LoRA Metadata Field + */ + LoRAMetadataField: { + /** @description LoRA model to load */ + model: components["schemas"]["ModelIdentifierField"]; + /** + * Weight + * @description The weight at which the LoRA is applied to each model + */ + weight: number; + }; + /** + * LoRA Name Grabber + * @description Outputs the LoRA's name as a string. Use a Lora Selector node as input. + */ + LoRANameGrabberInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRA + * @description The LoRA whose name you wish to grab. Get this from a LoRA Selector node. + * @default null + */ + lora?: components["schemas"]["LoRAField"] | null; + /** + * type + * @default lora_name_grabber_invocation + * @constant + */ + type: "lora_name_grabber_invocation"; + }; + /** + * LoRANameGrabberOutput + * @description LoRA Name Grabber output + */ + LoRANameGrabberOutput: { + /** + * Name + * @description Name of the LoRA + */ + name: string; + /** + * type + * @default lora_name_grabber_output + * @constant + */ + type: "lora_name_grabber_output"; + }; + /** + * LoRARecallParameter + * @description LoRA configuration for recall + */ + LoRARecallParameter: { + /** + * Model Name + * @description The name of the LoRA model + */ + model_name: string; + /** + * Weight + * @description The weight for the LoRA + * @default 0.75 + */ + weight?: number; + /** + * Is Enabled + * @description Whether the LoRA is enabled + * @default true + */ + is_enabled?: boolean; + }; + /** + * Select LoRA + * @description Selects a LoRA model and weight. + */ + LoRASelectorInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRA + * @description LoRA model to load + * @default null + */ + lora?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 + */ + weight?: number; + /** + * type + * @default lora_selector + * @constant + */ + type: "lora_selector"; + }; + /** + * LoRASelectorOutput + * @description Model loader output + */ + LoRASelectorOutput: { + /** + * LoRA + * @description LoRA model and weight + */ + lora: components["schemas"]["LoRAField"]; + /** + * type + * @default lora_selector_output + * @constant + */ + type: "lora_selector_output"; + }; + /** + * LoRA Toggle + * @description Allows boolean selection between two separate LoRA inputs + */ + LoRAToggleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Use Second + * @description Use 2nd Input + * @default false + */ + use_second?: boolean; + /** + * @description First LoRA Input + * @default null + */ + lora1?: components["schemas"]["LoRAField"] | null; + /** + * @description Second LoRA Input + * @default null + */ + lora2?: components["schemas"]["LoRAField"] | null; + /** + * type + * @default lora_toggle + * @constant + */ + type: "lora_toggle"; + }; + /** LoRAToggleOutput */ + LoRAToggleOutput: { + /** + * LoRA + * @description LoRA model and weight + */ + lora: components["schemas"]["LoRAField"]; + /** + * type + * @default lora_toggle_output + * @constant + */ + type: "lora_toggle_output"; + }; + /** LoRA_Diffusers_FLUX_Config */ + LoRA_Diffusers_FLUX_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; + }; + /** + * LoRA_Diffusers_Flux2_Config + * @description Model config for FLUX.2 (Klein) LoRA models in Diffusers format. + */ + LoRA_Diffusers_Flux2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** + * Base + * @default flux2 + * @constant + */ + base: "flux2"; + variant: components["schemas"]["Flux2VariantType"] | null; + }; + /** LoRA_Diffusers_SD1_Config */ + LoRA_Diffusers_SD1_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** + * Base + * @default sd-1 + * @constant + */ + base: "sd-1"; + }; + /** LoRA_Diffusers_SD2_Config */ + LoRA_Diffusers_SD2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; + }; + /** LoRA_Diffusers_SDXL_Config */ + LoRA_Diffusers_SDXL_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; + }; + /** + * LoRA_Diffusers_ZImage_Config + * @description Model config for Z-Image LoRA models in Diffusers format. + */ + LoRA_Diffusers_ZImage_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** + * Base + * @default z-image + * @constant + */ + base: "z-image"; + variant: components["schemas"]["ZImageVariantType"] | null; + }; + /** LoRA_LyCORIS_FLUX_Config */ + LoRA_LyCORIS_FLUX_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default lycoris + * @constant + */ + format: "lycoris"; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; + }; + /** + * LoRA_LyCORIS_Flux2_Config + * @description Model config for FLUX.2 (Klein) LoRA models in LyCORIS format. + */ + LoRA_LyCORIS_Flux2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default lycoris + * @constant + */ + format: "lycoris"; + /** + * Base + * @default flux2 + * @constant + */ + base: "flux2"; + variant: components["schemas"]["Flux2VariantType"] | null; + }; + /** LoRA_LyCORIS_SD1_Config */ + LoRA_LyCORIS_SD1_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default lycoris + * @constant + */ + format: "lycoris"; + /** + * Base + * @default sd-1 + * @constant + */ + base: "sd-1"; + }; + /** LoRA_LyCORIS_SD2_Config */ + LoRA_LyCORIS_SD2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default lycoris + * @constant + */ + format: "lycoris"; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; + }; + /** LoRA_LyCORIS_SDXL_Config */ + LoRA_LyCORIS_SDXL_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default lycoris + * @constant + */ + format: "lycoris"; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; + }; + /** + * LoRA_LyCORIS_ZImage_Config + * @description Model config for Z-Image LoRA models in LyCORIS format. + */ + LoRA_LyCORIS_ZImage_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default lycoris + * @constant + */ + format: "lycoris"; + /** + * Base + * @default z-image + * @constant + */ + base: "z-image"; + variant: components["schemas"]["ZImageVariantType"] | null; + }; + /** LoRA_OMI_FLUX_Config */ + LoRA_OMI_FLUX_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default omi + * @constant + */ + format: "omi"; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; + }; + /** LoRA_OMI_SDXL_Config */ + LoRA_OMI_SDXL_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default omi + * @constant + */ + format: "omi"; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; + }; + /** + * Load All Text Files In Folder + * @description Loads all text files in a folder and its subfolders recursively, returning them as a Collection of strings + */ + LoadAllTextFilesInFolderInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * File Extension + * @description Only files with the given extension will be loaded. For example: json + * @default null + */ + extension_to_match?: string | null; + /** + * Folder Path + * @description The path to load files from. + * @default null + */ + folder_path?: string | null; + /** + * type + * @default load_all_text_files_in_folder_output + * @constant + */ + type: "load_all_text_files_in_folder_output"; + }; + /** + * LoadAllTextFilesInFolderOutput + * @description Load Text Files In Folder Output + */ + LoadAllTextFilesInFolderOutput: { + /** Results */ + result: string[]; + /** + * type + * @default load_all_text_files_in_folder_output + * @constant + */ + type: "load_all_text_files_in_folder_output"; + }; + /** + * Load Aperture Image + * @description Loads a grayscale aperture image from the presets directory. + */ + LoadApertureImageInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Aperture Image + * @description The aperture image to load from the 'aperture_images' folder + * @default 5_point_star.png + * @enum {string} + */ + aperture_image?: "5_point_star.png" | "cross.png" | "diag_arrow.png" | "diffraction.png" | "heart.png"; + /** + * type + * @default load_aperture_image + * @constant + */ + type: "load_aperture_image"; + }; + /** + * Load Text File to String + * @description Loads a text from a provided path and outputs it as a string + */ + LoadTextFileToStringInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false + */ + use_cache?: boolean; + /** + * Path + * @description The full path to the text file. + * @default null + */ + file_path?: string | null; + /** + * type + * @default load_text_file_to_string_invocation + * @constant + */ + type: "load_text_file_to_string_invocation"; + }; + /** + * LoadTextFileToStringOutput + * @description Load Text File To String Output + */ + LoadTextFileToStringOutput: { + /** Result */ + result: string; + /** + * type + * @default load_text_file_to_string_output + * @constant + */ + type: "load_text_file_to_string_output"; + }; + /** + * Load Video Frame + * @description Load a specific frame from an MP4 video and provide it as output. + */ + LoadVideoFrameInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Video Path + * @description The path to the MP4 video file + * @default null + */ + video_path?: string | null; + /** + * Frame Number + * @description The frame number to load + * @default 1 + */ + frame_number?: number; + /** + * type + * @default load_video_frame + * @constant + */ + type: "load_video_frame"; + }; + /** + * LocalModelSource + * @description A local file or directory path. + */ + LocalModelSource: { + /** Path */ + path: string; + /** + * Inplace + * @default false + */ + inplace?: boolean | null; + /** + * @description discriminator enum property added by openapi-typescript + * @enum {string} + */ + type: "local"; + }; + /** + * LogLevel + * @enum {integer} + */ + LogLevel: 0 | 10 | 20 | 30 | 40 | 50; + /** + * LoginRequest + * @description Request body for user login. + */ + LoginRequest: { + /** + * Email + * @description User email address + */ + email: string; + /** + * Password + * @description User password + */ + password: string; + /** + * Remember Me + * @description Whether to extend session duration + * @default false + */ + remember_me?: boolean; + }; + /** + * LoginResponse + * @description Response from successful login. + */ + LoginResponse: { + /** + * Token + * @description JWT access token + */ + token: string; + /** @description User information */ + user: components["schemas"]["UserDTO"]; + /** + * Expires In + * @description Token expiration time in seconds + */ + expires_in: number; + }; + /** + * LogoutResponse + * @description Response from logout. + */ + LogoutResponse: { + /** + * Success + * @description Whether logout was successful + */ + success: boolean; + }; + /** + * Lookup LoRA Collection Triggers + * @description Retrieves trigger words from the Model Manager for all LoRAs in the collection + */ + LookupLoRACollectionTriggersInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRA Collection + * @description The LoRAs to look up + * @default null + */ + lora_collection?: components["schemas"]["LoRAField"][] | null; + /** + * type + * @default lookup_lora_collection_triggers_invocation + * @constant + */ + type: "lookup_lora_collection_triggers_invocation"; + }; + /** + * LookupLoRACollectionTriggersOutput + * @description Lookup LoRA Collection Triggers Output + */ + LookupLoRACollectionTriggersOutput: { + /** + * Trigger Words + * @description A collection of all the LoRAs trigger words + */ + trigger_words: string[]; + /** + * type + * @default lookup_lora_collection_triggers_output + * @constant + */ + type: "lookup_lora_collection_triggers_output"; + }; + /** + * Lookup LoRA Triggers + * @description Retrieves a LoRA's trigger words from the Model Manager + */ + LookupLoRATriggersInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * LoRA + * @description The LoRA to look up + * @default null + */ + lora?: components["schemas"]["LoRAField"] | null; + /** + * type + * @default lookup_lora_triggers_invocation + * @constant + */ + type: "lookup_lora_triggers_invocation"; + }; + /** + * LookupLoRATriggersOutput + * @description Lookup LoRA Triggers Output + */ + LookupLoRATriggersOutput: { + /** + * Trigger Words + * @description A collection of the LoRAs trigger words + */ + trigger_words: string[]; + /** + * type + * @default lookup_lora_triggers_output + * @constant + */ + type: "lookup_lora_triggers_output"; + }; + /** + * Lookup Table from File + * @description Loads a lookup table from a YAML file + */ + LookupTableFromFileInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * File Path + * @description Path to lookup table YAML file + * @default null + */ + file_path?: string | null; + /** + * type + * @default lookup_table_from_file + * @constant + */ + type: "lookup_table_from_file"; + }; + /** + * LookupTableOutput + * @description Base class for invocations that output a JSON lookup table + */ + LookupTableOutput: { + /** + * Lookups + * @description The output lookup table + */ + lookups: string; + /** + * type + * @default lookups_output + * @constant + */ + type: "lookups_output"; + }; + /** + * Lookups Entry from Prompt + * @description Creates a lookup table of a single heading->value + */ + LookupsEntryFromPromptInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Heading + * @description Heading for the lookup table entry + * @default null + */ + heading?: string | null; + /** + * Lookup + * @description The entry to place under Heading in the lookup table + * @default + */ + lookup?: string; + /** + * type + * @default lookup_from_prompt + * @constant + */ + type: "lookup_from_prompt"; + }; + /** LoraModelDefaultSettings */ + LoraModelDefaultSettings: { + /** + * Weight + * @description Default weight for this model + */ + weight?: number | null; + }; + /** + * LoRA To String + * @description Converts an lora Model to a JSONString + */ + LoraToStringInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description LoRA model to load + * @default null + */ + model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default lora_to_string + * @constant + */ + type: "lora_to_string"; + }; + /** MDControlListOutput */ + MDControlListOutput: { + /** + * ControlNet-List + * @description ControlNet(s) to apply + */ + control_list: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + /** + * type + * @default md_control_list_output + * @constant + */ + type: "md_control_list_output"; + }; + /** MDIPAdapterListOutput */ + MDIPAdapterListOutput: { + /** + * IP-Adapter-List + * @description IP-Adapter to apply + */ + ip_adapter_list: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + /** + * type + * @default md_ip_adapter_list_output + * @constant + */ + type: "md_ip_adapter_list_output"; + }; + /** MDT2IAdapterListOutput */ + MDT2IAdapterListOutput: { + /** + * T2I Adapter-List + * @description T2I-Adapter(s) to apply + */ + t2i_adapter_list: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + /** + * type + * @default md_ip_adapters_output + * @constant + */ + type: "md_ip_adapters_output"; + }; + /** + * MLSD Detection + * @description Generates an line segment map using MLSD. + */ + MLSDDetectionInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to process + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Score Threshold + * @description The threshold used to score points when determining line segments + * @default 0.1 + */ + score_threshold?: number; + /** + * Distance Threshold + * @description Threshold for including a line segment - lines shorter than this distance will be discarded + * @default 20 + */ + distance_threshold?: number; + /** + * type + * @default mlsd_detection + * @constant + */ + type: "mlsd_detection"; + }; + /** MainModelDefaultSettings */ + MainModelDefaultSettings: { + /** + * Vae + * @description Default VAE for this model (model key) + */ + vae?: string | null; + /** + * Vae Precision + * @description Default VAE precision for this model + */ + vae_precision?: ("fp16" | "fp32") | null; + /** + * Scheduler + * @description Default scheduler for this model + */ + scheduler?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; + /** + * Steps + * @description Default number of steps for this model + */ + steps?: number | null; + /** + * Cfg Scale + * @description Default CFG Scale for this model + */ + cfg_scale?: number | null; + /** + * Cfg Rescale Multiplier + * @description Default CFG Rescale Multiplier for this model + */ + cfg_rescale_multiplier?: number | null; + /** + * Width + * @description Default width for this model + */ + width?: number | null; + /** + * Height + * @description Default height for this model + */ + height?: number | null; + /** + * Guidance + * @description Default Guidance for this model + */ + guidance?: number | null; + /** + * Cpu Only + * @description Whether this model should run on CPU only + */ + cpu_only?: boolean | null; + }; + /** + * Main Model Input + * @description Loads a main model from an input, outputting its submodels. + */ + MainModelLoaderInputInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Main model (UNet, VAE, CLIP) to load + * @default null + */ + model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default main_model_loader_input + * @constant + */ + type: "main_model_loader_input"; + }; + /** + * Main Model - SD1.5, SD2 + * @description Loads a main model, outputting its submodels. + */ + MainModelLoaderInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Main model (UNet, VAE, CLIP) to load + * @default null + */ + model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default main_model_loader + * @constant + */ + type: "main_model_loader"; + }; + /** + * Main Model To String + * @description Converts a Main Model to a JSONString + */ + MainModelToStringInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Main model (UNet, VAE, CLIP) to load + * @default null + */ + model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default main_model_to_string + * @constant + */ + type: "main_model_to_string"; + }; + /** + * Main_BnBNF4_FLUX_Config + * @description Model config for main checkpoint models. + */ + Main_BnBNF4_FLUX_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; + /** + * Format + * @default bnb_quantized_nf4b + * @constant + */ + format: "bnb_quantized_nf4b"; + variant: components["schemas"]["FluxVariantType"]; + }; + /** + * Main_Checkpoint_FLUX_Config + * @description Model config for main checkpoint models. + */ + Main_Checkpoint_FLUX_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; + variant: components["schemas"]["FluxVariantType"]; + }; + /** + * Main_Checkpoint_Flux2_Config + * @description Model config for FLUX.2 checkpoint models (e.g. Klein). + */ + Main_Checkpoint_Flux2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + /** + * Base + * @default flux2 + * @constant + */ + base: "flux2"; + variant: components["schemas"]["Flux2VariantType"]; + }; + /** Main_Checkpoint_SD1_Config */ + Main_Checkpoint_SD1_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; + /** + * Base + * @default sd-1 + * @constant + */ + base: "sd-1"; + }; + /** Main_Checkpoint_SD2_Config */ + Main_Checkpoint_SD2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; + }; + /** Main_Checkpoint_SDXLRefiner_Config */ + Main_Checkpoint_SDXLRefiner_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; + /** + * Base + * @default sdxl-refiner + * @constant + */ + base: "sdxl-refiner"; + }; + /** Main_Checkpoint_SDXL_Config */ + Main_Checkpoint_SDXL_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; + }; + /** + * Main_Checkpoint_ZImage_Config + * @description Model config for Z-Image single-file checkpoint models (safetensors, etc). + */ + Main_Checkpoint_ZImage_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Base + * @default z-image + * @constant + */ + base: "z-image"; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; + variant: components["schemas"]["ZImageVariantType"]; + }; + /** Main_Diffusers_CogView4_Config */ + Main_Diffusers_CogView4_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + /** + * Base + * @default cogview4 + * @constant + */ + base: "cogview4"; + }; + /** + * Main_Diffusers_FLUX_Config + * @description Model config for FLUX.1 models in diffusers format. */ - InvocationProgressEvent: { + Main_Diffusers_FLUX_Config: { /** - * Timestamp - * @description The timestamp of the event + * Key + * @description A unique key for this model. */ - timestamp: number; + key: string; /** - * Queue Id - * @description The ID of the queue + * Hash + * @description The hash of the model file(s). */ - queue_id: string; + hash: string; /** - * Item Id - * @description The ID of the queue item + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - item_id: number; + path: string; /** - * Batch Id - * @description The ID of the queue batch + * File Size + * @description The size of the model in bytes. */ - batch_id: string; + file_size: number; /** - * Origin - * @description The origin of the queue item - * @default null + * Name + * @description Name of the model. */ - origin: string | null; + name: string; /** - * Destination - * @description The destination of the queue item - * @default null + * Description + * @description Model description */ - destination: string | null; + description: string | null; /** - * User Id - * @description The ID of the user who created the queue item - * @default system + * Source + * @description The original source of the model (path, URL or repo_id). */ - user_id: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Session Id - * @description The ID of the session (aka graph execution state) + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - session_id: string; + source_api_response: string | null; /** - * Invocation - * @description The ID of the invocation + * Cover Image + * @description Url for image to preview model */ - invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + cover_image: string | null; /** - * Invocation Source Id - * @description The ID of the prepared invocation's source node + * Type + * @default main + * @constant */ - invocation_source_id: string; + type: "main"; /** - * Message - * @description A message to display + * Trigger Phrases + * @description Set of trigger phrases for this model */ - message: string; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Percentage - * @description The percentage of the progress (omit to indicate indeterminate progress) - * @default null + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + /** + * Base + * @default flux + * @constant + */ + base: "flux"; + variant: components["schemas"]["FluxVariantType"]; + }; + /** + * Main_Diffusers_Flux2_Config + * @description Model config for FLUX.2 models in diffusers format (e.g. FLUX.2 Klein). + */ + Main_Diffusers_Flux2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + /** + * Base + * @default flux2 + * @constant */ - percentage: number | null; + base: "flux2"; + variant: components["schemas"]["Flux2VariantType"]; + }; + /** Main_Diffusers_SD1_Config */ + Main_Diffusers_SD1_Config: { /** - * @description An image representing the current state of the progress - * @default null + * Key + * @description A unique key for this model. */ - image: components["schemas"]["ProgressImage"] | null; - }; - /** - * InvocationStartedEvent - * @description Event model for invocation_started - */ - InvocationStartedEvent: { + key: string; /** - * Timestamp - * @description The timestamp of the event + * Hash + * @description The hash of the model file(s). */ - timestamp: number; + hash: string; /** - * Queue Id - * @description The ID of the queue + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - queue_id: string; + path: string; /** - * Item Id - * @description The ID of the queue item + * File Size + * @description The size of the model in bytes. */ - item_id: number; + file_size: number; /** - * Batch Id - * @description The ID of the queue batch + * Name + * @description Name of the model. */ - batch_id: string; + name: string; /** - * Origin - * @description The origin of the queue item - * @default null + * Description + * @description Model description */ - origin: string | null; + description: string | null; /** - * Destination - * @description The destination of the queue item - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - destination: string | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * User Id - * @description The ID of the user who created the queue item - * @default system + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - user_id: string; + source_api_response: string | null; /** - * Session Id - * @description The ID of the session (aka graph execution state) + * Cover Image + * @description Url for image to preview model */ - session_id: string; + cover_image: string | null; /** - * Invocation - * @description The ID of the invocation + * Type + * @default main + * @constant */ - invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + type: "main"; /** - * Invocation Source Id - * @description The ID of the prepared invocation's source node + * Trigger Phrases + * @description Set of trigger phrases for this model */ - invocation_source_id: string; - }; - /** - * InvokeAIAppConfig - * @description Invoke's global app configuration. - * - * Typically, you won't need to interact with this class directly. Instead, use the `get_config` function from `invokeai.app.services.config` to get a singleton config object. - * - * Attributes: - * host: IP address to bind to. Use `0.0.0.0` to serve to your local network. - * port: Port to bind to. - * allow_origins: Allowed CORS origins. - * allow_credentials: Allow CORS credentials. - * allow_methods: Methods allowed for CORS. - * allow_headers: Headers allowed for CORS. - * ssl_certfile: SSL certificate file for HTTPS. See https://www.uvicorn.org/settings/#https. - * ssl_keyfile: SSL key file for HTTPS. See https://www.uvicorn.org/settings/#https. - * log_tokenization: Enable logging of parsed prompt tokens. - * patchmatch: Enable patchmatch inpaint code. - * models_dir: Path to the models directory. - * convert_cache_dir: Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). - * download_cache_dir: Path to the directory that contains dynamically downloaded models. - * legacy_conf_dir: Path to directory of legacy checkpoint config files. - * db_dir: Path to InvokeAI databases directory. - * outputs_dir: Path to directory for outputs. - * custom_nodes_dir: Path to directory for custom nodes. - * style_presets_dir: Path to directory for style presets. - * workflow_thumbnails_dir: Path to directory for workflow thumbnails. - * log_handlers: Log handler. Valid options are "console", "file=", "syslog=path|address:host:port", "http=". - * log_format: Log format. Use "plain" for text-only, "color" for colorized output, "legacy" for 2.3-style logging and "syslog" for syslog-style.
Valid values: `plain`, `color`, `syslog`, `legacy` - * log_level: Emit logging messages at this level or higher.
Valid values: `debug`, `info`, `warning`, `error`, `critical` - * log_sql: Log SQL queries. `log_level` must be `debug` for this to do anything. Extremely verbose. - * log_level_network: Log level for network-related messages. 'info' and 'debug' are very verbose.
Valid values: `debug`, `info`, `warning`, `error`, `critical` - * use_memory_db: Use in-memory database. Useful for development. - * dev_reload: Automatically reload when Python sources are changed. Does not reload node definitions. - * profile_graphs: Enable graph profiling using `cProfile`. - * profile_prefix: An optional prefix for profile output files. - * profiles_dir: Path to profiles output directory. - * max_cache_ram_gb: The maximum amount of CPU RAM to use for model caching in GB. If unset, the limit will be configured based on the available RAM. In most cases, it is recommended to leave this unset. - * max_cache_vram_gb: The amount of VRAM to use for model caching in GB. If unset, the limit will be configured based on the available VRAM and the device_working_mem_gb. In most cases, it is recommended to leave this unset. - * log_memory_usage: If True, a memory snapshot will be captured before and after every model cache operation, and the result will be logged (at debug level). There is a time cost to capturing the memory snapshots, so it is recommended to only enable this feature if you are actively inspecting the model cache's behaviour. - * model_cache_keep_alive_min: How long to keep models in cache after last use, in minutes. A value of 0 (the default) means models are kept in cache indefinitely. If no model generations occur within the timeout period, the model cache is cleared using the same logic as the 'Clear Model Cache' button. - * device_working_mem_gb: The amount of working memory to keep available on the compute device (in GB). Has no effect if running on CPU. If you are experiencing OOM errors, try increasing this value. - * enable_partial_loading: Enable partial loading of models. This enables models to run with reduced VRAM requirements (at the cost of slower speed) by streaming the model from RAM to VRAM as its used. In some edge cases, partial loading can cause models to run more slowly if they were previously being fully loaded into VRAM. - * keep_ram_copy_of_weights: Whether to keep a full RAM copy of a model's weights when the model is loaded in VRAM. Keeping a RAM copy increases average RAM usage, but speeds up model switching and LoRA patching (assuming there is sufficient RAM). Set this to False if RAM pressure is consistently high. - * ram: DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_ram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. - * vram: DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_vram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. - * lazy_offload: DEPRECATED: This setting is no longer used. Lazy-offloading is enabled by default. This config setting will be removed once the new model cache behavior is stable. - * pytorch_cuda_alloc_conf: Configure the Torch CUDA memory allocator. This will impact peak reserved VRAM usage and performance. Setting to "backend:cudaMallocAsync" works well on many systems. The optimal configuration is highly dependent on the system configuration (device type, VRAM, CUDA driver version, etc.), so must be tuned experimentally. - * device: Preferred execution device. `auto` will choose the device depending on the hardware platform and the installed torch capabilities.
Valid values: `auto`, `cpu`, `cuda`, `mps`, `cuda:N` (where N is a device number) - * precision: Floating point precision. `float16` will consume half the memory of `float32` but produce slightly lower-quality images. The `auto` setting will guess the proper precision based on your video card and operating system.
Valid values: `auto`, `float16`, `bfloat16`, `float32` - * sequential_guidance: Whether to calculate guidance in serial instead of in parallel, lowering memory requirements. - * attention_type: Attention type.
Valid values: `auto`, `normal`, `xformers`, `sliced`, `torch-sdp` - * attention_slice_size: Slice size, valid when attention_type=="sliced".
Valid values: `auto`, `balanced`, `max`, `1`, `2`, `3`, `4`, `5`, `6`, `7`, `8` - * force_tiled_decode: Whether to enable tiled VAE decode (reduces memory consumption with some performance penalty). - * pil_compress_level: The compress_level setting of PIL.Image.save(), used for PNG encoding. All settings are lossless. 0 = no compression, 1 = fastest with slightly larger filesize, 9 = slowest with smallest filesize. 1 is typically the best setting. - * max_queue_size: Maximum number of items in the session queue. - * clear_queue_on_startup: Empties session queue on startup. - * allow_nodes: List of nodes to allow. Omit to allow all. - * deny_nodes: List of nodes to deny. Omit to deny none. - * node_cache_size: How many cached nodes to keep in memory. - * hashing_algorithm: Model hashing algorthim for model installs. 'blake3_multi' is best for SSDs. 'blake3_single' is best for spinning disk HDDs. 'random' disables hashing, instead assigning a UUID to models. Useful when using a memory db to reduce model installation time, or if you don't care about storing stable hashes for models. Alternatively, any other hashlib algorithm is accepted, though these are not nearly as performant as blake3.
Valid values: `blake3_multi`, `blake3_single`, `random`, `md5`, `sha1`, `sha224`, `sha256`, `sha384`, `sha512`, `blake2b`, `blake2s`, `sha3_224`, `sha3_256`, `sha3_384`, `sha3_512`, `shake_128`, `shake_256` - * remote_api_tokens: List of regular expression and token pairs used when downloading models from URLs. The download URL is tested against the regex, and if it matches, the token is provided in as a Bearer token. - * scan_models_on_startup: Scan the models directory on startup, registering orphaned models. This is typically only used in conjunction with `use_memory_db` for testing purposes. - * unsafe_disable_picklescan: UNSAFE. Disable the picklescan security check during model installation. Recommended only for development and testing purposes. This will allow arbitrary code execution during model installation, so should never be used in production. - * allow_unknown_models: Allow installation of models that we are unable to identify. If enabled, models will be marked as `unknown` in the database, and will not have any metadata associated with them. If disabled, unknown models will be rejected during installation. - * multiuser: Enable multiuser support. When disabled, the application runs in single-user mode using a default system account with administrator privileges. When enabled, requires user authentication and authorization. - * strict_password_checking: Enforce strict password requirements. When True, passwords must contain uppercase, lowercase, and numbers. When False (default), any password is accepted but its strength (weak/moderate/strong) is reported to the user. - */ - InvokeAIAppConfig: { + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Schema Version - * @description Schema version of the config file. This is not a user-configurable setting. - * @default 4.0.2 + * Format + * @default diffusers + * @constant */ - schema_version?: string; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * Legacy Models Yaml Path - * @description Path to the legacy models.yaml file. This is not a user-configurable setting. + * Base + * @default sd-1 + * @constant */ - legacy_models_yaml_path?: string | null; + base: "sd-1"; + }; + /** Main_Diffusers_SD2_Config */ + Main_Diffusers_SD2_Config: { /** - * Host - * @description IP address to bind to. Use `0.0.0.0` to serve to your local network. - * @default 127.0.0.1 + * Key + * @description A unique key for this model. */ - host?: string; + key: string; /** - * Port - * @description Port to bind to. - * @default 9090 + * Hash + * @description The hash of the model file(s). */ - port?: number; + hash: string; /** - * Allow Origins - * @description Allowed CORS origins. - * @default [] + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - allow_origins?: string[]; + path: string; /** - * Allow Credentials - * @description Allow CORS credentials. - * @default true + * File Size + * @description The size of the model in bytes. */ - allow_credentials?: boolean; + file_size: number; /** - * Allow Methods - * @description Methods allowed for CORS. - * @default [ - * "*" - * ] + * Name + * @description Name of the model. */ - allow_methods?: string[]; + name: string; /** - * Allow Headers - * @description Headers allowed for CORS. - * @default [ - * "*" - * ] + * Description + * @description Model description */ - allow_headers?: string[]; + description: string | null; /** - * Ssl Certfile - * @description SSL certificate file for HTTPS. See https://www.uvicorn.org/settings/#https. + * Source + * @description The original source of the model (path, URL or repo_id). */ - ssl_certfile?: string | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Ssl Keyfile - * @description SSL key file for HTTPS. See https://www.uvicorn.org/settings/#https. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - ssl_keyfile?: string | null; + source_api_response: string | null; /** - * Log Tokenization - * @description Enable logging of parsed prompt tokens. - * @default false + * Cover Image + * @description Url for image to preview model */ - log_tokenization?: boolean; + cover_image: string | null; /** - * Patchmatch - * @description Enable patchmatch inpaint code. - * @default true + * Type + * @default main + * @constant */ - patchmatch?: boolean; + type: "main"; /** - * Models Dir - * Format: path - * @description Path to the models directory. - * @default models + * Trigger Phrases + * @description Set of trigger phrases for this model */ - models_dir?: string; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Convert Cache Dir - * Format: path - * @description Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). - * @default models/.convert_cache + * Format + * @default diffusers + * @constant */ - convert_cache_dir?: string; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * Download Cache Dir - * Format: path - * @description Path to the directory that contains dynamically downloaded models. - * @default models/.download_cache + * Base + * @default sd-2 + * @constant */ - download_cache_dir?: string; + base: "sd-2"; + }; + /** Main_Diffusers_SD3_Config */ + Main_Diffusers_SD3_Config: { /** - * Legacy Conf Dir - * Format: path - * @description Path to directory of legacy checkpoint config files. - * @default configs + * Key + * @description A unique key for this model. */ - legacy_conf_dir?: string; + key: string; /** - * Db Dir - * Format: path - * @description Path to InvokeAI databases directory. - * @default databases + * Hash + * @description The hash of the model file(s). */ - db_dir?: string; + hash: string; /** - * Outputs Dir - * Format: path - * @description Path to directory for outputs. - * @default outputs + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - outputs_dir?: string; + path: string; /** - * Custom Nodes Dir - * Format: path - * @description Path to directory for custom nodes. - * @default nodes + * File Size + * @description The size of the model in bytes. */ - custom_nodes_dir?: string; + file_size: number; /** - * Style Presets Dir - * Format: path - * @description Path to directory for style presets. - * @default style_presets + * Name + * @description Name of the model. */ - style_presets_dir?: string; + name: string; /** - * Workflow Thumbnails Dir - * Format: path - * @description Path to directory for workflow thumbnails. - * @default workflow_thumbnails + * Description + * @description Model description */ - workflow_thumbnails_dir?: string; + description: string | null; /** - * Log Handlers - * @description Log handler. Valid options are "console", "file=", "syslog=path|address:host:port", "http=". - * @default [ - * "console" - * ] + * Source + * @description The original source of the model (path, URL or repo_id). */ - log_handlers?: string[]; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Log Format - * @description Log format. Use "plain" for text-only, "color" for colorized output, "legacy" for 2.3-style logging and "syslog" for syslog-style. - * @default color - * @enum {string} + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - log_format?: "plain" | "color" | "syslog" | "legacy"; + source_api_response: string | null; /** - * Log Level - * @description Emit logging messages at this level or higher. - * @default info - * @enum {string} + * Cover Image + * @description Url for image to preview model */ - log_level?: "debug" | "info" | "warning" | "error" | "critical"; + cover_image: string | null; /** - * Log Sql - * @description Log SQL queries. `log_level` must be `debug` for this to do anything. Extremely verbose. - * @default false + * Type + * @default main + * @constant */ - log_sql?: boolean; + type: "main"; /** - * Log Level Network - * @description Log level for network-related messages. 'info' and 'debug' are very verbose. - * @default warning - * @enum {string} + * Trigger Phrases + * @description Set of trigger phrases for this model */ - log_level_network?: "debug" | "info" | "warning" | "error" | "critical"; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Use Memory Db - * @description Use in-memory database. Useful for development. - * @default false + * Format + * @default diffusers + * @constant */ - use_memory_db?: boolean; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Dev Reload - * @description Automatically reload when Python sources are changed. Does not reload node definitions. - * @default false + * Base + * @default sd-3 + * @constant */ - dev_reload?: boolean; + base: "sd-3"; /** - * Profile Graphs - * @description Enable graph profiling using `cProfile`. - * @default false + * Submodels + * @description Loadable submodels in this model */ - profile_graphs?: boolean; + submodels: { + [key: string]: components["schemas"]["SubmodelDefinition"]; + } | null; + }; + /** Main_Diffusers_SDXLRefiner_Config */ + Main_Diffusers_SDXLRefiner_Config: { /** - * Profile Prefix - * @description An optional prefix for profile output files. + * Key + * @description A unique key for this model. */ - profile_prefix?: string | null; + key: string; /** - * Profiles Dir - * Format: path - * @description Path to profiles output directory. - * @default profiles + * Hash + * @description The hash of the model file(s). */ - profiles_dir?: string; + hash: string; /** - * Max Cache Ram Gb - * @description The maximum amount of CPU RAM to use for model caching in GB. If unset, the limit will be configured based on the available RAM. In most cases, it is recommended to leave this unset. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - max_cache_ram_gb?: number | null; + path: string; /** - * Max Cache Vram Gb - * @description The amount of VRAM to use for model caching in GB. If unset, the limit will be configured based on the available VRAM and the device_working_mem_gb. In most cases, it is recommended to leave this unset. + * File Size + * @description The size of the model in bytes. */ - max_cache_vram_gb?: number | null; + file_size: number; /** - * Log Memory Usage - * @description If True, a memory snapshot will be captured before and after every model cache operation, and the result will be logged (at debug level). There is a time cost to capturing the memory snapshots, so it is recommended to only enable this feature if you are actively inspecting the model cache's behaviour. - * @default false + * Name + * @description Name of the model. */ - log_memory_usage?: boolean; + name: string; /** - * Model Cache Keep Alive Min - * @description How long to keep models in cache after last use, in minutes. A value of 0 (the default) means models are kept in cache indefinitely. If no model generations occur within the timeout period, the model cache is cleared using the same logic as the 'Clear Model Cache' button. - * @default 0 + * Description + * @description Model description */ - model_cache_keep_alive_min?: number; + description: string | null; /** - * Device Working Mem Gb - * @description The amount of working memory to keep available on the compute device (in GB). Has no effect if running on CPU. If you are experiencing OOM errors, try increasing this value. - * @default 3 + * Source + * @description The original source of the model (path, URL or repo_id). */ - device_working_mem_gb?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Enable Partial Loading - * @description Enable partial loading of models. This enables models to run with reduced VRAM requirements (at the cost of slower speed) by streaming the model from RAM to VRAM as its used. In some edge cases, partial loading can cause models to run more slowly if they were previously being fully loaded into VRAM. - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - enable_partial_loading?: boolean; + source_api_response: string | null; /** - * Keep Ram Copy Of Weights - * @description Whether to keep a full RAM copy of a model's weights when the model is loaded in VRAM. Keeping a RAM copy increases average RAM usage, but speeds up model switching and LoRA patching (assuming there is sufficient RAM). Set this to False if RAM pressure is consistently high. - * @default true + * Cover Image + * @description Url for image to preview model */ - keep_ram_copy_of_weights?: boolean; + cover_image: string | null; /** - * Ram - * @description DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_ram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. + * Type + * @default main + * @constant */ - ram?: number | null; + type: "main"; /** - * Vram - * @description DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_vram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. + * Trigger Phrases + * @description Set of trigger phrases for this model */ - vram?: number | null; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Lazy Offload - * @description DEPRECATED: This setting is no longer used. Lazy-offloading is enabled by default. This config setting will be removed once the new model cache behavior is stable. - * @default true + * Format + * @default diffusers + * @constant */ - lazy_offload?: boolean; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * Pytorch Cuda Alloc Conf - * @description Configure the Torch CUDA memory allocator. This will impact peak reserved VRAM usage and performance. Setting to "backend:cudaMallocAsync" works well on many systems. The optimal configuration is highly dependent on the system configuration (device type, VRAM, CUDA driver version, etc.), so must be tuned experimentally. + * Base + * @default sdxl-refiner + * @constant */ - pytorch_cuda_alloc_conf?: string | null; + base: "sdxl-refiner"; + }; + /** Main_Diffusers_SDXL_Config */ + Main_Diffusers_SDXL_Config: { /** - * Device - * @description Preferred execution device. `auto` will choose the device depending on the hardware platform and the installed torch capabilities.
Valid values: `auto`, `cpu`, `cuda`, `mps`, `cuda:N` (where N is a device number) - * @default auto + * Key + * @description A unique key for this model. */ - device?: string; + key: string; /** - * Precision - * @description Floating point precision. `float16` will consume half the memory of `float32` but produce slightly lower-quality images. The `auto` setting will guess the proper precision based on your video card and operating system. - * @default auto - * @enum {string} + * Hash + * @description The hash of the model file(s). */ - precision?: "auto" | "float16" | "bfloat16" | "float32"; + hash: string; /** - * Sequential Guidance - * @description Whether to calculate guidance in serial instead of in parallel, lowering memory requirements. - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - sequential_guidance?: boolean; + path: string; /** - * Attention Type - * @description Attention type. - * @default auto - * @enum {string} + * File Size + * @description The size of the model in bytes. */ - attention_type?: "auto" | "normal" | "xformers" | "sliced" | "torch-sdp"; + file_size: number; /** - * Attention Slice Size - * @description Slice size, valid when attention_type=="sliced". - * @default auto - * @enum {unknown} + * Name + * @description Name of the model. */ - attention_slice_size?: "auto" | "balanced" | "max" | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8; + name: string; /** - * Force Tiled Decode - * @description Whether to enable tiled VAE decode (reduces memory consumption with some performance penalty). - * @default false + * Description + * @description Model description */ - force_tiled_decode?: boolean; + description: string | null; /** - * Pil Compress Level - * @description The compress_level setting of PIL.Image.save(), used for PNG encoding. All settings are lossless. 0 = no compression, 1 = fastest with slightly larger filesize, 9 = slowest with smallest filesize. 1 is typically the best setting. - * @default 1 + * Source + * @description The original source of the model (path, URL or repo_id). */ - pil_compress_level?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Max Queue Size - * @description Maximum number of items in the session queue. - * @default 10000 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - max_queue_size?: number; + source_api_response: string | null; /** - * Clear Queue On Startup - * @description Empties session queue on startup. - * @default false + * Cover Image + * @description Url for image to preview model */ - clear_queue_on_startup?: boolean; + cover_image: string | null; /** - * Allow Nodes - * @description List of nodes to allow. Omit to allow all. + * Type + * @default main + * @constant */ - allow_nodes?: string[] | null; + type: "main"; /** - * Deny Nodes - * @description List of nodes to deny. Omit to deny none. + * Trigger Phrases + * @description Set of trigger phrases for this model */ - deny_nodes?: string[] | null; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Node Cache Size - * @description How many cached nodes to keep in memory. - * @default 512 + * Format + * @default diffusers + * @constant */ - node_cache_size?: number; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * Hashing Algorithm - * @description Model hashing algorthim for model installs. 'blake3_multi' is best for SSDs. 'blake3_single' is best for spinning disk HDDs. 'random' disables hashing, instead assigning a UUID to models. Useful when using a memory db to reduce model installation time, or if you don't care about storing stable hashes for models. Alternatively, any other hashlib algorithm is accepted, though these are not nearly as performant as blake3. - * @default blake3_single - * @enum {string} + * Base + * @default sdxl + * @constant */ - hashing_algorithm?: "blake3_multi" | "blake3_single" | "random" | "md5" | "sha1" | "sha224" | "sha256" | "sha384" | "sha512" | "blake2b" | "blake2s" | "sha3_224" | "sha3_256" | "sha3_384" | "sha3_512" | "shake_128" | "shake_256"; + base: "sdxl"; + }; + /** + * Main_Diffusers_ZImage_Config + * @description Model config for Z-Image diffusers models (Z-Image-Turbo, Z-Image-Base). + */ + Main_Diffusers_ZImage_Config: { /** - * Remote Api Tokens - * @description List of regular expression and token pairs used when downloading models from URLs. The download URL is tested against the regex, and if it matches, the token is provided in as a Bearer token. + * Key + * @description A unique key for this model. */ - remote_api_tokens?: components["schemas"]["URLRegexTokenPair"][] | null; + key: string; /** - * Scan Models On Startup - * @description Scan the models directory on startup, registering orphaned models. This is typically only used in conjunction with `use_memory_db` for testing purposes. - * @default false + * Hash + * @description The hash of the model file(s). */ - scan_models_on_startup?: boolean; + hash: string; /** - * Unsafe Disable Picklescan - * @description UNSAFE. Disable the picklescan security check during model installation. Recommended only for development and testing purposes. This will allow arbitrary code execution during model installation, so should never be used in production. - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - unsafe_disable_picklescan?: boolean; + path: string; /** - * Allow Unknown Models - * @description Allow installation of models that we are unable to identify. If enabled, models will be marked as `unknown` in the database, and will not have any metadata associated with them. If disabled, unknown models will be rejected during installation. - * @default true + * File Size + * @description The size of the model in bytes. */ - allow_unknown_models?: boolean; + file_size: number; /** - * Multiuser - * @description Enable multiuser support. When disabled, the application runs in single-user mode using a default system account with administrator privileges. When enabled, requires user authentication and authorization. - * @default false + * Name + * @description Name of the model. */ - multiuser?: boolean; + name: string; /** - * Strict Password Checking - * @description Enforce strict password requirements. When True, passwords must contain uppercase, lowercase, and numbers. When False (default), any password is accepted but its strength (weak/moderate/strong) is reported to the user. - * @default false + * Description + * @description Model description */ - strict_password_checking?: boolean; - }; - /** - * InvokeAIAppConfigWithSetFields - * @description InvokeAI App Config with model fields set - */ - InvokeAIAppConfigWithSetFields: { + description: string | null; /** - * Set Fields - * @description The set fields + * Source + * @description The original source of the model (path, URL or repo_id). */ - set_fields: string[]; - /** @description The InvokeAI App Config */ - config: components["schemas"]["InvokeAIAppConfig"]; - }; - /** - * Adjust Image Hue Plus - * @description Adjusts the Hue of an image by rotating it in the selected color space. Originally created by @dwringer - */ - InvokeAdjustImageHuePlusInvocation: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * @description The board to save the image to - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - board?: components["schemas"]["BoardField"] | null; + source_api_response: string | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Cover Image + * @description Url for image to preview model */ - metadata?: components["schemas"]["MetadataField"] | null; + cover_image: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Type + * @default main + * @constant */ - id: string; + type: "main"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Trigger Phrases + * @description Set of trigger phrases for this model */ - is_intermediate?: boolean; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Format + * @default diffusers + * @constant */ - use_cache?: boolean; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * @description The image to adjust - * @default null + * Base + * @default z-image + * @constant */ - image?: components["schemas"]["ImageField"] | null; + base: "z-image"; + variant: components["schemas"]["ZImageVariantType"]; + }; + /** + * Main_GGUF_FLUX_Config + * @description Model config for main checkpoint models. + */ + Main_GGUF_FLUX_Config: { /** - * Space - * @description Color space in which to rotate hue by polar coords (*: non-invertible) - * @default HSV / HSL / RGB - * @enum {string} + * Key + * @description A unique key for this model. */ - space?: "HSV / HSL / RGB" | "Okhsl" | "Okhsv" | "*Oklch / Oklab" | "*LCh / CIELab" | "*UPLab (w/CIELab_to_UPLab.icc)"; + key: string; /** - * Degrees - * @description Degrees by which to rotate image hue - * @default 0 + * Hash + * @description The hash of the model file(s). */ - degrees?: number; + hash: string; /** - * Preserve Lightness - * @description Whether to preserve CIELAB lightness values - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - preserve_lightness?: boolean; + path: string; /** - * Ok Adaptive Gamut - * @description Higher preserves chroma at the expense of lightness (Oklab) - * @default 0.05 + * File Size + * @description The size of the model in bytes. */ - ok_adaptive_gamut?: number; + file_size: number; /** - * Ok High Precision - * @description Use more steps in computing gamut (Oklab/Okhsv/Okhsl) - * @default true + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description */ - ok_high_precision?: boolean; + description: string | null; /** - * type - * @default invokeai_img_hue_adjust_plus - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "invokeai_img_hue_adjust_plus"; - }; - /** - * Equivalent Achromatic Lightness - * @description Calculate Equivalent Achromatic Lightness from image. Originally created by @dwringer - */ - InvokeEquivalentAchromaticLightnessInvocation: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * @description The board to save the image to - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - board?: components["schemas"]["BoardField"] | null; + source_api_response: string | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Cover Image + * @description Url for image to preview model */ - metadata?: components["schemas"]["MetadataField"] | null; + cover_image: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Type + * @default main + * @constant */ - id: string; + type: "main"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Trigger Phrases + * @description Set of trigger phrases for this model */ - is_intermediate?: boolean; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Config Path + * @description Path to the config for this model, if any. */ - use_cache?: boolean; + config_path: string | null; /** - * @description Image from which to get channel - * @default null + * Base + * @default flux + * @constant */ - image?: components["schemas"]["ImageField"] | null; + base: "flux"; /** - * type - * @default invokeai_ealightness + * Format + * @default gguf_quantized * @constant */ - type: "invokeai_ealightness"; + format: "gguf_quantized"; + variant: components["schemas"]["FluxVariantType"]; }; /** - * Image Layer Blend - * @description Blend two images together, with optional opacity, mask, and blend modes. Originally created by @dwringer + * Main_GGUF_Flux2_Config + * @description Model config for GGUF-quantized FLUX.2 checkpoint models (e.g. Klein). */ - InvokeImageBlendInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + Main_GGUF_Flux2_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * @description The top image to blend - * @default null + * File Size + * @description The size of the model in bytes. */ - layer_upper?: components["schemas"]["ImageField"] | null; + file_size: number; /** - * Blend Mode - * @description Available blend modes - * @default Normal - * @enum {string} + * Name + * @description Name of the model. */ - blend_mode?: "Normal" | "Lighten Only" | "Darken Only" | "Lighten Only (EAL)" | "Darken Only (EAL)" | "Hue" | "Saturation" | "Color" | "Luminosity" | "Linear Dodge (Add)" | "Subtract" | "Multiply" | "Divide" | "Screen" | "Overlay" | "Linear Burn" | "Difference" | "Hard Light" | "Soft Light" | "Vivid Light" | "Linear Light" | "Color Burn" | "Color Dodge"; + name: string; /** - * Opacity - * @description Desired opacity of the upper layer - * @default 1 + * Description + * @description Model description */ - opacity?: number; + description: string | null; /** - * @description Optional mask, used to restrict areas from blending - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - mask?: components["schemas"]["ImageField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Fit To Width - * @description Scale upper layer to fit base width - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - fit_to_width?: boolean; + source_api_response: string | null; /** - * Fit To Height - * @description Scale upper layer to fit base height - * @default true + * Cover Image + * @description Url for image to preview model */ - fit_to_height?: boolean; + cover_image: string | null; /** - * @description The bottom image to blend - * @default null + * Type + * @default main + * @constant */ - layer_base?: components["schemas"]["ImageField"] | null; + type: "main"; /** - * Color Space - * @description Available color spaces for blend computations - * @default RGB - * @enum {string} + * Trigger Phrases + * @description Set of trigger phrases for this model */ - color_space?: "RGB" | "Linear RGB" | "HSL (RGB)" | "HSV (RGB)" | "Okhsl" | "Okhsv" | "Oklch (Oklab)" | "LCh (CIELab)"; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Adaptive Gamut - * @description Adaptive gamut clipping (0=off). Higher prioritizes chroma over lightness - * @default 0 + * Config Path + * @description Path to the config for this model, if any. */ - adaptive_gamut?: number; + config_path: string | null; /** - * High Precision - * @description Use more steps in computing gamut when possible - * @default true + * Base + * @default flux2 + * @constant */ - high_precision?: boolean; + base: "flux2"; /** - * type - * @default invokeai_img_blend + * Format + * @default gguf_quantized * @constant */ - type: "invokeai_img_blend"; + format: "gguf_quantized"; + variant: components["schemas"]["Flux2VariantType"]; }; /** - * Image Compositor - * @description Removes backdrop from subject image then overlays subject on background image. Originally created by @dwringer + * Main_GGUF_ZImage_Config + * @description Model config for GGUF-quantized Z-Image transformer models. */ - InvokeImageCompositorInvocation: { + Main_GGUF_ZImage_Config: { /** - * @description The board to save the image to - * @default null + * Key + * @description A unique key for this model. */ - board?: components["schemas"]["BoardField"] | null; + key: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Hash + * @description The hash of the model file(s). */ - metadata?: components["schemas"]["MetadataField"] | null; + hash: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - id: string; + path: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * File Size + * @description The size of the model in bytes. */ - is_intermediate?: boolean; + file_size: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Name + * @description Name of the model. */ - use_cache?: boolean; + name: string; /** - * @description Image of the subject on a plain monochrome background - * @default null + * Description + * @description Model description */ - image_subject?: components["schemas"]["ImageField"] | null; + description: string | null; /** - * @description Image of a background scene - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - image_background?: components["schemas"]["ImageField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Chroma Key - * @description Can be empty for corner flood select, or CSS-3 color or tuple - * @default + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - chroma_key?: string; + source_api_response: string | null; /** - * Threshold - * @description Subject isolation flood-fill threshold - * @default 50 + * Cover Image + * @description Url for image to preview model */ - threshold?: number; - /** - * Fill X - * @description Scale base subject image to fit background width - * @default false + cover_image: string | null; + /** + * Type + * @default main + * @constant */ - fill_x?: boolean; + type: "main"; /** - * Fill Y - * @description Scale base subject image to fit background height - * @default true + * Trigger Phrases + * @description Set of trigger phrases for this model */ - fill_y?: boolean; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * X Offset - * @description x-offset for the subject - * @default 0 + * Config Path + * @description Path to the config for this model, if any. */ - x_offset?: number; + config_path: string | null; /** - * Y Offset - * @description y-offset for the subject - * @default 0 + * Base + * @default z-image + * @constant */ - y_offset?: number; + base: "z-image"; /** - * type - * @default invokeai_img_composite + * Format + * @default gguf_quantized * @constant */ - type: "invokeai_img_composite"; + format: "gguf_quantized"; + variant: components["schemas"]["ZImageVariantType"]; }; /** - * Image Dilate or Erode - * @description Dilate (expand) or erode (contract) an image. Originally created by @dwringer + * Combine Masks + * @description Combine two masks together by multiplying them using `PIL.ImageChops.multiply()`. */ - InvokeImageDilateOrErodeInvocation: { + MaskCombineInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -15212,47 +28661,27 @@ export type components = { */ use_cache?: boolean; /** - * @description The image from which to create a mask + * @description The first mask to combine * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * Lightness Only - * @description If true, only applies to image lightness (CIELa*b*) - * @default false - */ - lightness_only?: boolean; - /** - * Radius W - * @description Width (in pixels) by which to dilate(expand) or erode (contract) the image - * @default 4 - */ - radius_w?: number; - /** - * Radius H - * @description Height (in pixels) by which to dilate(expand) or erode (contract) the image - * @default 4 - */ - radius_h?: number; + mask1?: components["schemas"]["ImageField"] | null; /** - * Mode - * @description How to operate on the image - * @default Dilate - * @enum {string} + * @description The second image to combine + * @default null */ - mode?: "Dilate" | "Erode"; + mask2?: components["schemas"]["ImageField"] | null; /** * type - * @default invokeai_img_dilate_erode + * @default mask_combine * @constant */ - type: "invokeai_img_dilate_erode"; + type: "mask_combine"; }; /** - * Enhance Image - * @description Applies processing from PIL's ImageEnhance module. Originally created by @dwringer + * Mask Edge + * @description Applies an edge mask to an image */ - InvokeImageEnhanceInvocation: { + MaskEdgeInvocation: { /** * @description The board to save the image to * @default null @@ -15281,52 +28710,46 @@ export type components = { */ use_cache?: boolean; /** - * @description The image for which to apply processing + * @description The image to apply the mask to * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Invert - * @description Whether to invert the image colors - * @default false - */ - invert?: boolean; - /** - * Color - * @description Color enhancement factor - * @default 1 + * Edge Size + * @description The size of the edge + * @default null */ - color?: number; + edge_size?: number | null; /** - * Contrast - * @description Contrast enhancement factor - * @default 1 + * Edge Blur + * @description The amount of blur on the edge + * @default null */ - contrast?: number; + edge_blur?: number | null; /** - * Brightness - * @description Brightness enhancement factor - * @default 1 + * Low Threshold + * @description First threshold for the hysteresis procedure in Canny edge detection + * @default null */ - brightness?: number; + low_threshold?: number | null; /** - * Sharpness - * @description Sharpness enhancement factor - * @default 1 + * High Threshold + * @description Second threshold for the hysteresis procedure in Canny edge detection + * @default null */ - sharpness?: number; + high_threshold?: number | null; /** * type - * @default invokeai_img_enhance + * @default mask_edge * @constant */ - type: "invokeai_img_enhance"; + type: "mask_edge"; }; /** - * Image Value Thresholds - * @description Clip image to pure black/white past specified thresholds. Originally created by @dwringer + * Mask from Alpha + * @description Extracts the alpha channel of an image as a mask. */ - InvokeImageValueThresholdsInvocation: { + MaskFromAlphaInvocation: { /** * @description The board to save the image to * @default null @@ -15355,68 +28778,38 @@ export type components = { */ use_cache?: boolean; /** - * @description The image from which to create a mask + * @description The image to create the mask from * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Invert Output - * @description Make light areas dark and vice versa - * @default false - */ - invert_output?: boolean; - /** - * Renormalize Values - * @description Rescale remaining values from minimum to maximum - * @default false - */ - renormalize_values?: boolean; - /** - * Lightness Only - * @description If true, only applies to image lightness (CIELa*b*) + * Invert + * @description Whether or not to invert the mask * @default false */ - lightness_only?: boolean; - /** - * Threshold Upper - * @description Threshold above which will be set to full value - * @default 0.5 - */ - threshold_upper?: number; - /** - * Threshold Lower - * @description Threshold below which will be set to minimum value - * @default 0.5 - */ - threshold_lower?: number; + invert?: boolean; /** * type - * @default invokeai_img_val_thresholds + * @default tomask * @constant */ - type: "invokeai_img_val_thresholds"; + type: "tomask"; }; /** - * ItemIdsResult - * @description Response containing ordered item ids with metadata for optimistic updates. + * Mask from Segmented Image + * @description Generate a mask for a particular color in an ID Map */ - ItemIdsResult: { + MaskFromIDInvocation: { /** - * Item Ids - * @description Ordered list of item ids + * @description The board to save the image to + * @default null */ - item_ids: number[]; + board?: components["schemas"]["BoardField"] | null; /** - * Total Count - * @description Total number of queue items matching the query + * @description Optional metadata to be saved with the image + * @default null */ - total_count: number; - }; - /** - * IterateInvocation - * @description Iterates over a list of items - */ - IterateInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -15435,57 +28828,63 @@ export type components = { */ use_cache?: boolean; /** - * Collection - * @description The list of items to iterate over - * @default [] + * @description The image to create the mask from + * @default null */ - collection?: unknown[]; + image?: components["schemas"]["ImageField"] | null; /** - * Index - * @description The index, will be provided on executed iterators - * @default 0 + * @description ID color to mask + * @default null */ - index?: number; + color?: components["schemas"]["ColorField"] | null; + /** + * Threshold + * @description Threshold for color detection + * @default 100 + */ + threshold?: number; + /** + * Invert + * @description Whether or not to invert the mask + * @default false + */ + invert?: boolean; /** * type - * @default iterate + * @default mask_from_id * @constant */ - type: "iterate"; + type: "mask_from_id"; }; /** - * IterateInvocationOutput - * @description Used to connect iteration outputs. Will be expanded to a specific output. + * MaskOutput + * @description A torch mask tensor. */ - IterateInvocationOutput: { - /** - * Collection Item - * @description The item being iterated over - */ - item: unknown; + MaskOutput: { + /** @description The mask. */ + mask: components["schemas"]["TensorField"]; /** - * Index - * @description The index of the item + * Width + * @description The width of the mask in pixels. */ - index: number; + width: number; /** - * Total - * @description The total number of items + * Height + * @description The height of the mask in pixels. */ - total: number; + height: number; /** * type - * @default iterate_output + * @default mask_output * @constant */ - type: "iterate_output"; + type: "mask_output"; }; - JsonValue: unknown; /** - * LaMa Infill - * @description Infills transparent areas of an image using the LaMa model + * Tensor Mask to Image + * @description Convert a mask tensor to an image. */ - LaMaInfillInvocation: { + MaskTensorToImageInvocation: { /** * @description The board to save the image to * @default null @@ -15514,22 +28913,32 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The mask tensor to convert. * @default null */ - image?: components["schemas"]["ImageField"] | null; + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default infill_lama + * @default tensor_mask_to_image * @constant */ - type: "infill_lama"; + type: "tensor_mask_to_image"; }; /** - * Latents Collection Primitive - * @description A collection of latents tensor primitive values + * Match Histogram (YCbCr) + * @description Match a histogram from one image to another using YCbCr color space */ - LatentsCollectionInvocation: { + MatchHistogramInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -15548,57 +28957,49 @@ export type components = { */ use_cache?: boolean; /** - * Collection - * @description The collection of latents tensors + * @description The image to receive the histogram * @default null */ - collection?: components["schemas"]["LatentsField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * type - * @default latents_collection - * @constant + * @description The reference image with the source histogram + * @default null */ - type: "latents_collection"; - }; - /** - * LatentsCollectionOutput - * @description Base class for nodes that output a collection of latents tensors - */ - LatentsCollectionOutput: { + reference_image?: components["schemas"]["ImageField"] | null; /** - * Collection - * @description Latents tensor + * Match Luminance Only + * @description Only transfer the luminance + * @default false */ - collection: components["schemas"]["LatentsField"][]; + match_luminance_only?: boolean; + /** + * Output Grayscale + * @description Convert output image to grayscale + * @default false + */ + output_grayscale?: boolean; /** * type - * @default latents_collection_output + * @default match_histogram * @constant */ - type: "latents_collection_output"; + type: "match_histogram"; }; /** - * LatentsField - * @description A latents tensor primitive field + * Match Histogram LAB + * @description Match a histogram from one image to another using Lab color space */ - LatentsField: { + MatchHistogramLabInvocation: { /** - * Latents Name - * @description The name of the latents + * @description The board to save the image to + * @default null */ - latents_name: string; + board?: components["schemas"]["BoardField"] | null; /** - * Seed - * @description Seed used to generate this latents + * @description Optional metadata to be saved with the image * @default null */ - seed?: number | null; - }; - /** - * Latents Primitive - * @description A latents tensor primitive value - */ - LatentsInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -15617,72 +29018,148 @@ export type components = { */ use_cache?: boolean; /** - * @description The latents tensor + * @description The image to receive the histogram * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + image?: components["schemas"]["ImageField"] | null; + /** + * @description The reference image with the source histogram + * @default null + */ + reference_image?: components["schemas"]["ImageField"] | null; + /** + * Match Luminance Only + * @description Only transfer the luminance/brightness channel + * @default false + */ + match_luminance_only?: boolean; + /** + * Output Grayscale + * @description Convert output image to grayscale + * @default false + */ + output_grayscale?: boolean; /** * type - * @default latents + * @default match_histogram_lab * @constant */ - type: "latents"; + type: "match_histogram_lab"; }; /** - * LatentsMetaOutput - * @description Latents + metadata + * MathEval + * @description Performs arbitrary user-defined calculations on x, y, z, and w variables */ - LatentsMetaOutput: { - /** @description Metadata Dict */ - metadata: components["schemas"]["MetadataField"]; + MathEvalInvocation: { /** - * type - * @default latents_meta_output - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - type: "latents_meta_output"; - /** @description Latents tensor */ - latents: components["schemas"]["LatentsField"]; + id: string; /** - * Width - * @description Width of output (px) + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - width: number; + is_intermediate?: boolean; /** - * Height - * @description Height of output (px) + * Use Cache + * @description Whether or not to use the cache + * @default true */ - height: number; + use_cache?: boolean; + /** + * X + * @description Input x + * @default 0 + */ + x?: number; + /** + * Y + * @description Input y + * @default 0 + */ + y?: number; + /** + * Z + * @description Input z + * @default 0 + */ + z?: number; + /** + * W + * @description Input w + * @default 0 + */ + w?: number; + /** + * Equation A + * @description A basic mathematical equation for a involving x, y, z, and w + * @default math.sin(x) + */ + equation_a?: string; + /** + * Equation B + * @description A basic mathematical equation for b involving x, y, z, and w + * @default 1+(x*y) + */ + equation_b?: string; + /** + * Equation C + * @description A basic mathematical equation for c involving x, y, z, and w + * @default + */ + equation_c?: string; + /** + * Equation D + * @description A basic mathematical equation for d involving x, y, z, and w + * @default + */ + equation_d?: string; + /** + * type + * @default math_eval + * @constant + */ + type: "math_eval"; }; /** - * LatentsOutput - * @description Base class for nodes that output a single latents tensor + * MathEvalOutput + * @description Base class for math output */ - LatentsOutput: { - /** @description Latents tensor */ - latents: components["schemas"]["LatentsField"]; + MathEvalOutput: { /** - * Width - * @description Width of output (px) + * A + * @description Output a */ - width: number; + a: number; /** - * Height - * @description Height of output (px) + * B + * @description Output b */ - height: number; + b: number; + /** + * C + * @description Output c + */ + c: number; + /** + * D + * @description Output d + */ + d: number; /** * type - * @default latents_output + * @default math_eval_output * @constant */ - type: "latents_output"; + type: "math_eval_output"; }; /** - * Latents to Image - SD1.5, SDXL - * @description Generates an image from latents. + * MediaPipe Face Detection + * @description Detects faces using MediaPipe. */ - LatentsToImageInvocation: { + MediaPipeFaceDetectionInvocation: { /** * @description The board to save the image to * @default null @@ -15711,55 +29188,34 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * @description VAE + * @description The image to process * @default null */ - vae?: components["schemas"]["VAEField"] | null; - /** - * Tiled - * @description Processing using overlapping tiles (reduce memory consumption) - * @default false - */ - tiled?: boolean; + image?: components["schemas"]["ImageField"] | null; /** - * Tile Size - * @description The tile size for VAE tiling in pixels (image space). If set to 0, the default tile size for the model will be used. Larger tile sizes generally produce better results at the cost of higher memory usage. - * @default 0 + * Max Faces + * @description Maximum number of faces to detect + * @default 1 */ - tile_size?: number; + max_faces?: number; /** - * Fp32 - * @description Whether or not to use full float32 precision - * @default false + * Min Confidence + * @description Minimum confidence for face detection + * @default 0.5 */ - fp32?: boolean; + min_confidence?: number; /** * type - * @default l2i + * @default mediapipe_face_detection * @constant */ - type: "l2i"; + type: "mediapipe_face_detection"; }; /** - * Lineart Anime Edge Detection - * @description Geneartes an edge map using the Lineart model. + * Merge LoRA Collections + * @description Merges two Collections of LoRAs into a single Collection. */ - LineartAnimeEdgeDetectionInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + MergeLoRACollectionsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -15778,32 +29234,46 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * Collection 1 + * @description A collection of LoRAs or a single LoRA from a LoRA Selector node. * @default null */ - image?: components["schemas"]["ImageField"] | null; + collection1?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; + /** + * Collection 2 + * @description A collection of LoRAs or a single LoRA from a LoRA Selector node. + * @default null + */ + collection2?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** * type - * @default lineart_anime_edge_detection + * @default merge_lora_collections_invocation * @constant */ - type: "lineart_anime_edge_detection"; + type: "merge_lora_collections_invocation"; }; /** - * Lineart Edge Detection - * @description Generates an edge map using the Lineart model. + * MergeLoRACollectionsOutput + * @description Join LoRA Collections output */ - LineartEdgeDetectionInvocation: { + MergeLoRACollectionsOutput: { /** - * @description The board to save the image to - * @default null + * Collection + * @description The merged collection */ - board?: components["schemas"]["BoardField"] | null; + collection: components["schemas"]["LoRAField"][]; /** - * @description Optional metadata to be saved with the image - * @default null + * type + * @default merge_lora_collections_output + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "merge_lora_collections_output"; + }; + /** + * Metadata Merge + * @description Merged a collection of MetadataDict into a single MetadataDict. + */ + MergeMetadataInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -15822,28 +29292,23 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Coarse - * @description Whether to use coarse mode - * @default false + * Collection + * @description Collection of Metadata + * @default null */ - coarse?: boolean; + collection?: components["schemas"]["MetadataField"][] | null; /** * type - * @default lineart_edge_detection + * @default merge_metadata * @constant */ - type: "lineart_edge_detection"; + type: "merge_metadata"; }; /** - * LLaVA OneVision VLLM - * @description Run a LLaVA OneVision VLLM model. + * Merge String Collections + * @description Merges two Collections of LoRAs into a single Collection. */ - LlavaOnevisionVllmInvocation: { + MergeStringCollectionsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -15862,113 +29327,111 @@ export type components = { */ use_cache?: boolean; /** - * Images - * @description Input image. + * Collection 1 + * @description A collection of strings or a single string. * @default null */ - images?: (components["schemas"]["ImageField"][] | components["schemas"]["ImageField"]) | null; - /** - * Prompt - * @description Input text prompt. - * @default - */ - prompt?: string; + collection1?: string | string[] | null; /** - * LLaVA Model Type - * @description The VLLM model to use + * Collection 2 + * @description A collection of strings or a single string. * @default null */ - vllm_model?: components["schemas"]["ModelIdentifierField"] | null; + collection2?: string | string[] | null; /** * type - * @default llava_onevision_vllm + * @default merge_string_collections_invocation * @constant */ - type: "llava_onevision_vllm"; + type: "merge_string_collections_invocation"; }; /** - * LlavaOnevision_Diffusers_Config - * @description Model config for Llava Onevision models. + * MergeStringCollectionsOutput + * @description Merge String Collections output */ - LlavaOnevision_Diffusers_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; + MergeStringCollectionsOutput: { /** - * Hash - * @description The hash of the model file(s). + * Collection + * @description The merged collection */ - hash: string; + collection: string[]; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * type + * @default merge_string_collections_output + * @constant */ - path: string; + type: "merge_string_collections_output"; + }; + /** + * Merge Tiles to Image + * @description Merge multiple tile images into a single image. + */ + MergeTilesToImageInvocation: { /** - * File Size - * @description The size of the model in bytes. + * @description The board to save the image to + * @default null */ - file_size: number; + board?: components["schemas"]["BoardField"] | null; /** - * Name - * @description Name of the model. + * @description Optional metadata to be saved with the image + * @default null */ - name: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Tiles With Images + * @description A list of tile images with tile properties. + * @default null */ - cover_image: string | null; + tiles_with_images?: components["schemas"]["TileWithImage"][] | null; /** - * Format - * @default diffusers - * @constant + * Blend Mode + * @description blending type Linear or Seam + * @default Seam + * @enum {string} */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + blend_mode?: "Linear" | "Seam"; /** - * Type - * @default llava_onevision - * @constant + * Blend Amount + * @description The amount to blend adjacent tiles in pixels. Must be <= the amount of overlap between adjacent tiles. + * @default 32 */ - type: "llava_onevision"; + blend_amount?: number; /** - * Base - * @default any + * type + * @default merge_tiles_to_image * @constant */ - base: "any"; - /** - * Cpu Only - * @description Whether this model should run on CPU only - */ - cpu_only: boolean | null; + type: "merge_tiles_to_image"; }; /** - * Apply LoRA Collection - SD1.5 - * @description Applies a collection of LoRAs to the provided UNet and CLIP models. + * MetadataField + * @description Pydantic model for metadata with custom root of type dict[str, Any]. + * Metadata is stored without a strict schema. */ - LoRACollectionLoader: { + MetadataField: Record; + /** + * Metadata Field Extractor + * @description Extracts the text value from an image's metadata given a key. + * Raises an error if the image has no metadata or if the value is not a string (nesting not permitted). + */ + MetadataFieldExtractorInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -15987,45 +29450,28 @@ export type components = { */ use_cache?: boolean; /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. - * @default null - */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; - /** - * UNet - * @description UNet (scheduler, LoRAs) + * @description The image to extract metadata from * @default null */ - unet?: components["schemas"]["UNetField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Key + * @description The key in the image's metadata to extract the value from * @default null */ - clip?: components["schemas"]["CLIPField"] | null; + key?: string | null; /** * type - * @default lora_collection_loader + * @default metadata_field_extractor * @constant */ - type: "lora_collection_loader"; - }; - /** LoRAField */ - LoRAField: { - /** @description Info to load lora model */ - lora: components["schemas"]["ModelIdentifierField"]; - /** - * Weight - * @description Weight to apply to lora model - */ - weight: number; + type: "metadata_field_extractor"; }; /** - * Apply LoRA - SD1.5 - * @description Apply selected lora to unet and text_encoder. + * Metadata From Image + * @description Used to create a core metadata item then Add/Update it to the provided metadata */ - LoRALoaderInvocation: { + MetadataFromImageInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -16044,101 +29490,70 @@ export type components = { */ use_cache?: boolean; /** - * LoRA - * @description LoRA model to load - * @default null - */ - lora?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 - */ - weight?: number; - /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @description The image to process * @default null */ - clip?: components["schemas"]["CLIPField"] | null; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default lora_loader + * @default metadata_from_image * @constant */ - type: "lora_loader"; + type: "metadata_from_image"; }; /** - * LoRALoaderOutput - * @description Model loader output + * Metadata + * @description Takes a MetadataItem or collection of MetadataItems and outputs a MetadataDict. */ - LoRALoaderOutput: { + MetadataInvocation: { /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - unet: components["schemas"]["UNetField"] | null; + id: string; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - clip: components["schemas"]["CLIPField"] | null; + is_intermediate?: boolean; /** - * type - * @default lora_loader_output - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - type: "lora_loader_output"; - }; - /** - * LoRAMetadataField - * @description LoRA Metadata Field - */ - LoRAMetadataField: { - /** @description LoRA model to load */ - model: components["schemas"]["ModelIdentifierField"]; + use_cache?: boolean; /** - * Weight - * @description The weight at which the LoRA is applied to each model + * Items + * @description A single metadata item or collection of metadata items + * @default null */ - weight: number; - }; - /** - * LoRARecallParameter - * @description LoRA configuration for recall - */ - LoRARecallParameter: { + items?: components["schemas"]["MetadataItemField"][] | components["schemas"]["MetadataItemField"] | null; /** - * Model Name - * @description The name of the LoRA model + * type + * @default metadata + * @constant */ - model_name: string; + type: "metadata"; + }; + /** MetadataItemField */ + MetadataItemField: { /** - * Weight - * @description The weight for the LoRA - * @default 0.75 + * Label + * @description Label for this metadata item */ - weight?: number; + label: string; /** - * Is Enabled - * @description Whether the LoRA is enabled - * @default true + * Value + * @description The value for this metadata item (may be any type) */ - is_enabled?: boolean; + value: unknown; }; /** - * Select LoRA - * @description Selects a LoRA model and weight. + * Metadata Item + * @description Used to create an arbitrary metadata item. Provide "label" and make a connection to "value" to store that data as the value. */ - LoRASelectorInvocation: { + MetadataItemInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -16157,3015 +29572,3331 @@ export type components = { */ use_cache?: boolean; /** - * LoRA - * @description LoRA model to load + * Label + * @description Label for this metadata item * @default null */ - lora?: components["schemas"]["ModelIdentifierField"] | null; + label?: string | null; /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 + * Value + * @description The value for this metadata item (may be any type) + * @default null */ - weight?: number; + value?: unknown | null; /** * type - * @default lora_selector + * @default metadata_item * @constant */ - type: "lora_selector"; + type: "metadata_item"; }; /** - * LoRASelectorOutput - * @description Model loader output + * Metadata Item Linked + * @description Used to Create/Add/Update a value into a metadata label */ - LoRASelectorOutput: { - /** - * LoRA - * @description LoRA model and weight - */ - lora: components["schemas"]["LoRAField"]; - /** - * type - * @default lora_selector_output - * @constant - */ - type: "lora_selector_output"; - }; - /** LoRA_Diffusers_FLUX_Config */ - LoRA_Diffusers_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; + MetadataItemLinkedInvocation: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * @description Optional metadata to be saved with the image + * @default null */ - path: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * File Size - * @description The size of the model in bytes. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - file_size: number; + id: string; /** - * Name - * @description Name of the model. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - name: string; + is_intermediate?: boolean; /** - * Description - * @description Model description + * Use Cache + * @description Whether or not to use the cache + * @default true */ - description: string | null; + use_cache?: boolean; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt" | "width" | "height" | "seed" | "cfg_scale" | "cfg_rescale_multiplier" | "steps" | "scheduler" | "clip_skip" | "model" | "vae" | "seamless_x" | "seamless_y" | "guidance" | "cfg_scale_start_step" | "cfg_scale_end_step"; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Custom Label + * @description Label for this metadata item + * @default null */ - source_api_response: string | null; + custom_label?: string | null; /** - * Cover Image - * @description Url for image to preview model + * Value + * @description The value for this metadata item (may be any type) + * @default null */ - cover_image: string | null; + value?: unknown | null; /** - * Type - * @default lora + * type + * @default metadata_item_linked * @constant */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + type: "metadata_item_linked"; + }; + /** + * MetadataItemOutput + * @description Metadata Item Output + */ + MetadataItemOutput: { + /** @description Metadata Item */ + item: components["schemas"]["MetadataItemField"]; /** - * Format - * @default diffusers + * type + * @default metadata_item_output * @constant */ - format: "diffusers"; + type: "metadata_item_output"; + }; + /** MetadataOutput */ + MetadataOutput: { + /** @description Metadata Dict */ + metadata: components["schemas"]["MetadataField"]; /** - * Base - * @default flux + * type + * @default metadata_output * @constant */ - base: "flux"; + type: "metadata_output"; }; /** - * LoRA_Diffusers_Flux2_Config - * @description Model config for FLUX.2 (Klein) LoRA models in Diffusers format. + * Metadata To Bool Collection + * @description Extracts a Boolean value Collection of a label from metadata */ - LoRA_Diffusers_Flux2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; + MetadataToBoolCollectionInvocation: { /** - * Hash - * @description The hash of the model file(s). + * @description Optional metadata to be saved with the image + * @default null */ - hash: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - description: string | null; + label?: "* CUSTOM LABEL *" | "seamless_x" | "seamless_y"; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Custom Label + * @description Label for this metadata item + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + custom_label?: string | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Default Value + * @description The default bool to use if not found in the metadata + * @default null */ - source_api_response: string | null; + default_value?: boolean[] | null; /** - * Cover Image - * @description Url for image to preview model + * type + * @default metadata_to_bool_collection + * @constant */ - cover_image: string | null; + type: "metadata_to_bool_collection"; + }; + /** + * Metadata To Bool + * @description Extracts a Boolean value of a label from metadata + */ + MetadataToBoolInvocation: { /** - * Type - * @default lora - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "lora"; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + id: string; /** - * Format - * @default diffusers - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - format: "diffusers"; + is_intermediate?: boolean; /** - * Base - * @default flux2 - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - base: "flux2"; - variant: components["schemas"]["Flux2VariantType"] | null; - }; - /** LoRA_Diffusers_SD1_Config */ - LoRA_Diffusers_SD1_Config: { + use_cache?: boolean; /** - * Key - * @description A unique key for this model. + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - key: string; + label?: "* CUSTOM LABEL *" | "seamless_x" | "seamless_y"; /** - * Hash - * @description The hash of the model file(s). + * Custom Label + * @description Label for this metadata item + * @default null */ - hash: string; + custom_label?: string | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Default Value + * @description The default bool to use if not found in the metadata + * @default null */ - path: string; + default_value?: boolean | null; /** - * File Size - * @description The size of the model in bytes. + * type + * @default metadata_to_bool + * @constant */ - file_size: number; + type: "metadata_to_bool"; + }; + /** + * Metadata To ControlNets + * @description Extracts a Controlnets value of a label from metadata + */ + MetadataToControlnetsInvocation: { /** - * Name - * @description Name of the model. + * @description Optional metadata to be saved with the image + * @default null */ - name: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * ControlNet-List + * @default null */ - cover_image: string | null; + control_list?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; /** - * Type - * @default lora + * type + * @default metadata_to_controlnets * @constant */ - type: "lora"; + type: "metadata_to_controlnets"; + }; + /** + * Metadata To Float Collection + * @description Extracts a Float value Collection of a label from metadata + */ + MetadataToFloatCollectionInvocation: { /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * @description Optional metadata to be saved with the image + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Format - * @default diffusers - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - format: "diffusers"; + id: string; /** - * Base - * @default sd-1 - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - base: "sd-1"; - }; - /** LoRA_Diffusers_SD2_Config */ - LoRA_Diffusers_SD2_Config: { + is_intermediate?: boolean; /** - * Key - * @description A unique key for this model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - key: string; + use_cache?: boolean; /** - * Hash - * @description The hash of the model file(s). + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - hash: string; + label?: "* CUSTOM LABEL *" | "cfg_scale" | "cfg_rescale_multiplier" | "guidance"; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Custom Label + * @description Label for this metadata item + * @default null */ - path: string; + custom_label?: string | null; /** - * File Size - * @description The size of the model in bytes. + * Default Value + * @description The default float to use if not found in the metadata + * @default null */ - file_size: number; + default_value?: number[] | null; /** - * Name - * @description Name of the model. + * type + * @default metadata_to_float_collection + * @constant */ - name: string; + type: "metadata_to_float_collection"; + }; + /** + * Metadata To Float + * @description Extracts a Float value of a label from metadata + */ + MetadataToFloatInvocation: { /** - * Description - * @description Model description + * @description Optional metadata to be saved with the image + * @default null */ - description: string | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Type - * @default lora - * @constant + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - type: "lora"; + label?: "* CUSTOM LABEL *" | "cfg_scale" | "cfg_rescale_multiplier" | "guidance"; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Custom Label + * @description Label for this metadata item + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + custom_label?: string | null; /** - * Format - * @default diffusers - * @constant + * Default Value + * @description The default float to use if not found in the metadata + * @default null */ - format: "diffusers"; + default_value?: number | null; /** - * Base - * @default sd-2 + * type + * @default metadata_to_float * @constant */ - base: "sd-2"; + type: "metadata_to_float"; }; - /** LoRA_Diffusers_SDXL_Config */ - LoRA_Diffusers_SDXL_Config: { + /** + * Metadata To IP-Adapters + * @description Extracts a IP-Adapters value of a label from metadata + */ + MetadataToIPAdaptersInvocation: { /** - * Key - * @description A unique key for this model. + * @description Optional metadata to be saved with the image + * @default null */ - key: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Hash - * @description The hash of the model file(s). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - hash: string; + id: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - path: string; + is_intermediate?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - file_size: number; + use_cache?: boolean; + /** + * IP-Adapter-List + * @description IP-Adapter to apply + * @default null + */ + ip_adapter_list?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * Name - * @description Name of the model. + * type + * @default metadata_to_ip_adapters + * @constant */ - name: string; + type: "metadata_to_ip_adapters"; + }; + /** + * Metadata To Integer Collection + * @description Extracts an integer value Collection of a label from metadata + */ + MetadataToIntegerCollectionInvocation: { /** - * Description - * @description Model description + * @description Optional metadata to be saved with the image + * @default null */ - description: string | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Type - * @default lora - * @constant + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - type: "lora"; + label?: "* CUSTOM LABEL *" | "width" | "height" | "seed" | "steps" | "clip_skip" | "cfg_scale_start_step" | "cfg_scale_end_step"; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Custom Label + * @description Label for this metadata item + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + custom_label?: string | null; /** - * Format - * @default diffusers - * @constant + * Default Value + * @description The default integer to use if not found in the metadata + * @default null */ - format: "diffusers"; + default_value?: number[] | null; /** - * Base - * @default sdxl + * type + * @default metadata_to_integer_collection * @constant */ - base: "sdxl"; + type: "metadata_to_integer_collection"; }; /** - * LoRA_Diffusers_ZImage_Config - * @description Model config for Z-Image LoRA models in Diffusers format. + * Metadata To Integer + * @description Extracts an integer value of a label from metadata */ - LoRA_Diffusers_ZImage_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; + MetadataToIntegerInvocation: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * @description Optional metadata to be saved with the image + * @default null */ - path: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * File Size - * @description The size of the model in bytes. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - file_size: number; + id: string; /** - * Name - * @description Name of the model. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - name: string; + is_intermediate?: boolean; /** - * Description - * @description Model description + * Use Cache + * @description Whether or not to use the cache + * @default true */ - description: string | null; + use_cache?: boolean; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + label?: "* CUSTOM LABEL *" | "width" | "height" | "seed" | "steps" | "clip_skip" | "cfg_scale_start_step" | "cfg_scale_end_step"; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Custom Label + * @description Label for this metadata item + * @default null */ - source_api_response: string | null; + custom_label?: string | null; /** - * Cover Image - * @description Url for image to preview model + * Default Value + * @description The default integer to use if not found in the metadata + * @default null */ - cover_image: string | null; + default_value?: number | null; /** - * Type - * @default lora + * type + * @default metadata_to_integer * @constant */ - type: "lora"; + type: "metadata_to_integer"; + }; + /** + * Metadata To LoRA Collection + * @description Extracts Lora(s) from metadata into a collection + */ + MetadataToLorasCollectionInvocation: { /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * @description Optional metadata to be saved with the image + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Format - * @default diffusers - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - format: "diffusers"; + id: string; /** - * Base - * @default z-image - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - base: "z-image"; - variant: components["schemas"]["ZImageVariantType"] | null; - }; - /** LoRA_LyCORIS_FLUX_Config */ - LoRA_LyCORIS_FLUX_Config: { + is_intermediate?: boolean; /** - * Key - * @description A unique key for this model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - key: string; + use_cache?: boolean; /** - * Hash - * @description The hash of the model file(s). + * Custom Label + * @description Label for this metadata item + * @default loras */ - hash: string; + custom_label?: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. + * @default [] */ - path: string; + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * File Size - * @description The size of the model in bytes. + * type + * @default metadata_to_lora_collection + * @constant */ - file_size: number; + type: "metadata_to_lora_collection"; + }; + /** + * MetadataToLorasCollectionOutput + * @description Model loader output + */ + MetadataToLorasCollectionOutput: { /** - * Name - * @description Name of the model. + * LoRAs + * @description Collection of LoRA model and weights */ - name: string; + lora: components["schemas"]["LoRAField"][]; /** - * Description - * @description Model description + * type + * @default metadata_to_lora_collection_output + * @constant */ - description: string | null; + type: "metadata_to_lora_collection_output"; + }; + /** + * Metadata To LoRAs + * @description Extracts a Loras value of a label from metadata + */ + MetadataToLorasInvocation: { /** - * Source - * @description The original source of the model (path, URL or repo_id). + * @description Optional metadata to be saved with the image + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source_api_response: string | null; + id: string; /** - * Cover Image - * @description Url for image to preview model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true */ - type: "lora"; + use_cache?: boolean; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + unet?: components["schemas"]["UNetField"] | null; /** - * Format - * @default lycoris - * @constant + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - format: "lycoris"; + clip?: components["schemas"]["CLIPField"] | null; /** - * Base - * @default flux + * type + * @default metadata_to_loras * @constant */ - base: "flux"; + type: "metadata_to_loras"; }; /** - * LoRA_LyCORIS_Flux2_Config - * @description Model config for FLUX.2 (Klein) LoRA models in LyCORIS format. + * Metadata To Model + * @description Extracts a Model value of a label from metadata */ - LoRA_LyCORIS_Flux2_Config: { + MetadataToModelInvocation: { /** - * Key - * @description A unique key for this model. + * @description Optional metadata to be saved with the image + * @default null */ - key: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Hash - * @description The hash of the model file(s). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - hash: string; + id: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - path: string; + is_intermediate?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - file_size: number; + use_cache?: boolean; /** - * Name - * @description Name of the model. + * Label + * @description Label for this metadata item + * @default model + * @enum {string} */ - name: string; + label?: "* CUSTOM LABEL *" | "model"; /** - * Description - * @description Model description + * Custom Label + * @description Label for this metadata item + * @default null */ - description: string | null; + custom_label?: string | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * @description The default model to use if not found in the metadata + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + default_value?: components["schemas"]["ModelIdentifierField"] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * type + * @default metadata_to_model + * @constant */ - source_api_response: string | null; + type: "metadata_to_model"; + }; + /** + * MetadataToModelOutput + * @description String to main model output + */ + MetadataToModelOutput: { /** - * Cover Image - * @description Url for image to preview model + * Model + * @description Main model (UNet, VAE, CLIP) to load */ - cover_image: string | null; + model: components["schemas"]["ModelIdentifierField"]; /** - * Type - * @default lora - * @constant + * Name + * @description Model Name */ - type: "lora"; + name: string; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * UNet + * @description UNet (scheduler, LoRAs) */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + unet: components["schemas"]["UNetField"]; /** - * Format - * @default lycoris - * @constant + * VAE + * @description VAE */ - format: "lycoris"; + vae: components["schemas"]["VAEField"]; /** - * Base - * @default flux2 + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + */ + clip: components["schemas"]["CLIPField"]; + /** + * type + * @default metadata_to_model_output * @constant */ - base: "flux2"; - variant: components["schemas"]["Flux2VariantType"] | null; + type: "metadata_to_model_output"; }; - /** LoRA_LyCORIS_SD1_Config */ - LoRA_LyCORIS_SD1_Config: { + /** + * Metadata To SDXL LoRAs + * @description Extracts a SDXL Loras value of a label from metadata + */ + MetadataToSDXLLorasInvocation: { /** - * Key - * @description A unique key for this model. + * @description Optional metadata to be saved with the image + * @default null */ - key: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Hash - * @description The hash of the model file(s). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - hash: string; + id: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - path: string; + is_intermediate?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - file_size: number; + use_cache?: boolean; /** - * Name - * @description Name of the model. + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - name: string; + unet?: components["schemas"]["UNetField"] | null; /** - * Description - * @description Model description + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - description: string | null; + clip?: components["schemas"]["CLIPField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + clip2?: components["schemas"]["CLIPField"] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * type + * @default metadata_to_sdlx_loras + * @constant */ - source_api_response: string | null; + type: "metadata_to_sdlx_loras"; + }; + /** + * Metadata To SDXL Model + * @description Extracts a SDXL Model value of a label from metadata + */ + MetadataToSDXLModelInvocation: { /** - * Cover Image - * @description Url for image to preview model + * @description Optional metadata to be saved with the image + * @default null */ - cover_image: string | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Type - * @default lora - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - type: "lora"; + id: string; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + is_intermediate?: boolean; /** - * Format - * @default lycoris - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - format: "lycoris"; + use_cache?: boolean; /** - * Base - * @default sd-1 - * @constant + * Label + * @description Label for this metadata item + * @default model + * @enum {string} */ - base: "sd-1"; - }; - /** LoRA_LyCORIS_SD2_Config */ - LoRA_LyCORIS_SD2_Config: { + label?: "* CUSTOM LABEL *" | "model"; /** - * Key - * @description A unique key for this model. + * Custom Label + * @description Label for this metadata item + * @default null */ - key: string; + custom_label?: string | null; /** - * Hash - * @description The hash of the model file(s). + * @description The default SDXL Model to use if not found in the metadata + * @default null */ - hash: string; + default_value?: components["schemas"]["ModelIdentifierField"] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * type + * @default metadata_to_sdxl_model + * @constant */ - path: string; + type: "metadata_to_sdxl_model"; + }; + /** + * MetadataToSDXLModelOutput + * @description String to SDXL main model output + */ + MetadataToSDXLModelOutput: { /** - * File Size - * @description The size of the model in bytes. + * Model + * @description Main model (UNet, VAE, CLIP) to load */ - file_size: number; + model: components["schemas"]["ModelIdentifierField"]; /** * Name - * @description Name of the model. + * @description Model Name */ name: string; /** - * Description - * @description Model description + * UNet + * @description UNet (scheduler, LoRAs) */ - description: string | null; + unet: components["schemas"]["UNetField"]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + clip: components["schemas"]["CLIPField"]; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - source_api_response: string | null; + clip2: components["schemas"]["CLIPField"]; /** - * Cover Image - * @description Url for image to preview model + * VAE + * @description VAE */ - cover_image: string | null; + vae: components["schemas"]["VAEField"]; /** - * Type - * @default lora + * type + * @default metadata_to_sdxl_model_output * @constant */ - type: "lora"; + type: "metadata_to_sdxl_model_output"; + }; + /** + * Metadata To Scheduler + * @description Extracts a Scheduler value of a label from metadata + */ + MetadataToSchedulerInvocation: { /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * @description Optional metadata to be saved with the image + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Format - * @default lycoris - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - format: "lycoris"; + id: string; /** - * Base - * @default sd-2 - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - base: "sd-2"; - }; - /** LoRA_LyCORIS_SDXL_Config */ - LoRA_LyCORIS_SDXL_Config: { + is_intermediate?: boolean; /** - * Key - * @description A unique key for this model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - key: string; + use_cache?: boolean; /** - * Hash - * @description The hash of the model file(s). + * Label + * @description Label for this metadata item + * @default scheduler + * @enum {string} */ - hash: string; + label?: "* CUSTOM LABEL *" | "scheduler"; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Custom Label + * @description Label for this metadata item + * @default null */ - path: string; + custom_label?: string | null; /** - * File Size - * @description The size of the model in bytes. + * Default Value + * @description The default scheduler to use if not found in the metadata + * @default euler + * @enum {string} */ - file_size: number; + default_value?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** - * Name - * @description Name of the model. + * type + * @default metadata_to_scheduler + * @constant */ - name: string; + type: "metadata_to_scheduler"; + }; + /** + * Metadata To String Collection + * @description Extracts a string collection value of a label from metadata + */ + MetadataToStringCollectionInvocation: { /** - * Description - * @description Model description + * @description Optional metadata to be saved with the image + * @default null */ - description: string | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Type - * @default lora - * @constant + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - type: "lora"; + label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt"; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Custom Label + * @description Label for this metadata item + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + custom_label?: string | null; /** - * Format - * @default lycoris - * @constant + * Default Value + * @description The default string collection to use if not found in the metadata + * @default null */ - format: "lycoris"; + default_value?: string[] | null; /** - * Base - * @default sdxl + * type + * @default metadata_to_string_collection * @constant */ - base: "sdxl"; + type: "metadata_to_string_collection"; }; /** - * LoRA_LyCORIS_ZImage_Config - * @description Model config for Z-Image LoRA models in LyCORIS format. + * Metadata To String + * @description Extracts a string value of a label from metadata */ - LoRA_LyCORIS_ZImage_Config: { + MetadataToStringInvocation: { /** - * Key - * @description A unique key for this model. + * @description Optional metadata to be saved with the image + * @default null */ - key: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Hash - * @description The hash of the model file(s). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - hash: string; + id: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - path: string; + is_intermediate?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - file_size: number; + use_cache?: boolean; + /** + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} + */ + label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt"; + /** + * Custom Label + * @description Label for this metadata item + * @default null + */ + custom_label?: string | null; + /** + * Default Value + * @description The default string to use if not found in the metadata + * @default null + */ + default_value?: string | null; + /** + * type + * @default metadata_to_string + * @constant + */ + type: "metadata_to_string"; + }; + /** + * Metadata To T2I-Adapters + * @description Extracts a T2I-Adapters value of a label from metadata + */ + MetadataToT2IAdaptersInvocation: { /** - * Name - * @description Name of the model. + * @description Optional metadata to be saved with the image + * @default null */ - name: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * T2I-Adapter + * @description IP-Adapter to apply + * @default null */ - cover_image: string | null; + t2i_adapter_list?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; /** - * Type - * @default lora + * type + * @default metadata_to_t2i_adapters * @constant */ - type: "lora"; + type: "metadata_to_t2i_adapters"; + }; + /** + * Metadata To VAE + * @description Extracts a VAE value of a label from metadata + */ + MetadataToVAEInvocation: { /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * @description Optional metadata to be saved with the image + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Format - * @default lycoris - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - format: "lycoris"; + id: string; /** - * Base - * @default z-image - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - base: "z-image"; - variant: components["schemas"]["ZImageVariantType"] | null; - }; - /** LoRA_OMI_FLUX_Config */ - LoRA_OMI_FLUX_Config: { + is_intermediate?: boolean; /** - * Key - * @description A unique key for this model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - key: string; + use_cache?: boolean; /** - * Hash - * @description The hash of the model file(s). + * Label + * @description Label for this metadata item + * @default vae + * @enum {string} */ - hash: string; + label?: "* CUSTOM LABEL *" | "vae"; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Custom Label + * @description Label for this metadata item + * @default null */ - path: string; + custom_label?: string | null; /** - * File Size - * @description The size of the model in bytes. + * @description The default VAE to use if not found in the metadata + * @default null */ - file_size: number; + default_value?: components["schemas"]["VAEField"] | null; /** - * Name - * @description Name of the model. + * type + * @default metadata_to_vae + * @constant */ - name: string; + type: "metadata_to_vae"; + }; + /** + * Minimum Overlap XYImage Tile Generator + * @description Cuts up an image into overlapping tiles and outputs a string representation of the tiles to use, taking the + * input overlap as a minimum + */ + MinimumOverlapXYTileGenerator: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * @description The input image + * @default null */ - cover_image: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Type - * @default lora - * @constant + * Tile Width + * @description x resolution of generation tile (must be a multiple of 8) + * @default 576 */ - type: "lora"; + tile_width?: number; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Tile Height + * @description y resolution of generation tile (must be a multiple of 8) + * @default 576 */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + tile_height?: number; /** - * Format - * @default omi - * @constant + * Min Overlap + * @description minimum tile overlap size (must be a multiple of 8) + * @default 128 */ - format: "omi"; + min_overlap?: number; /** - * Base - * @default flux + * Round To 8 + * @description Round outputs down to the nearest 8 (for pulling from a large noise field) + * @default false + */ + round_to_8?: boolean; + /** + * type + * @default minimum_overlap_xy_tile_generator * @constant */ - base: "flux"; + type: "minimum_overlap_xy_tile_generator"; }; - /** LoRA_OMI_SDXL_Config */ - LoRA_OMI_SDXL_Config: { + /** + * ModelFormat + * @description Storage format of model. + * @enum {string} + */ + ModelFormat: "omi" | "diffusers" | "checkpoint" | "lycoris" | "onnx" | "olive" | "embedding_file" | "embedding_folder" | "invokeai" | "t5_encoder" | "qwen3_encoder" | "bnb_quantized_int8b" | "bnb_quantized_nf4b" | "gguf_quantized" | "unknown"; + /** ModelIdentifierField */ + ModelIdentifierField: { /** * Key - * @description A unique key for this model. + * @description The model's unique key */ key: string; /** * Hash - * @description The hash of the model file(s). + * @description The model's BLAKE3 hash */ hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; /** * Name - * @description Name of the model. + * @description The model's name */ name: string; + /** @description The model's base model type */ + base: components["schemas"]["BaseModelType"]; + /** @description The model's type */ + type: components["schemas"]["ModelType"]; /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description The submodel to load, if this is a main model + * @default null */ - source_api_response: string | null; + submodel_type?: components["schemas"]["SubModelType"] | null; + }; + /** + * Any Model + * @description Selects any model, outputting it its identifier. Be careful with this one! The identifier will be accepted as + * input for any model, even if the model types don't match. If you connect this to a mismatched input, you'll get an + * error. + */ + ModelIdentifierInvocation: { /** - * Cover Image - * @description Url for image to preview model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - cover_image: string | null; + id: string; /** - * Type - * @default lora - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - type: "lora"; + is_intermediate?: boolean; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + use_cache?: boolean; /** - * Format - * @default omi - * @constant + * Model + * @description The model to select + * @default null */ - format: "omi"; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Base - * @default sdxl + * type + * @default model_identifier * @constant */ - base: "sdxl"; + type: "model_identifier"; }; /** - * LocalModelSource - * @description A local file or directory path. + * ModelIdentifierOutput + * @description Model identifier output */ - LocalModelSource: { - /** Path */ - path: string; + ModelIdentifierOutput: { /** - * Inplace - * @default false + * Model + * @description Model identifier */ - inplace?: boolean | null; + model: components["schemas"]["ModelIdentifierField"]; /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} + * type + * @default model_identifier_output + * @constant */ - type: "local"; + type: "model_identifier_output"; }; /** - * LogLevel - * @enum {integer} - */ - LogLevel: 0 | 10 | 20 | 30 | 40 | 50; - /** - * LoginRequest - * @description Request body for user login. + * ModelInstallCancelledEvent + * @description Event model for model_install_cancelled */ - LoginRequest: { + ModelInstallCancelledEvent: { /** - * Email - * @description User email address + * Timestamp + * @description The timestamp of the event */ - email: string; + timestamp: number; /** - * Password - * @description User password + * Id + * @description The ID of the install job */ - password: string; + id: number; /** - * Remember Me - * @description Whether to extend session duration - * @default false + * Source + * @description Source of the model; local path, repo_id or url */ - remember_me?: boolean; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; }; /** - * LoginResponse - * @description Response from successful login. + * ModelInstallCompleteEvent + * @description Event model for model_install_complete */ - LoginResponse: { + ModelInstallCompleteEvent: { /** - * Token - * @description JWT access token + * Timestamp + * @description The timestamp of the event */ - token: string; - /** @description User information */ - user: components["schemas"]["UserDTO"]; + timestamp: number; /** - * Expires In - * @description Token expiration time in seconds + * Id + * @description The ID of the install job */ - expires_in: number; - }; - /** - * LogoutResponse - * @description Response from logout. - */ - LogoutResponse: { + id: number; /** - * Success - * @description Whether logout was successful + * Source + * @description Source of the model; local path, repo_id or url */ - success: boolean; - }; - /** LoraModelDefaultSettings */ - LoraModelDefaultSettings: { + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * Weight - * @description Default weight for this model + * Key + * @description Model config record key */ - weight?: number | null; - }; - /** MDControlListOutput */ - MDControlListOutput: { + key: string; /** - * ControlNet-List - * @description ControlNet(s) to apply + * Total Bytes + * @description Size of the model (may be None for installation of a local path) */ - control_list: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + total_bytes: number | null; /** - * type - * @default md_control_list_output - * @constant + * Config + * @description The installed model's config */ - type: "md_control_list_output"; + config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; }; - /** MDIPAdapterListOutput */ - MDIPAdapterListOutput: { - /** - * IP-Adapter-List - * @description IP-Adapter to apply - */ - ip_adapter_list: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + /** + * ModelInstallDownloadProgressEvent + * @description Event model for model_install_download_progress + */ + ModelInstallDownloadProgressEvent: { /** - * type - * @default md_ip_adapter_list_output - * @constant + * Timestamp + * @description The timestamp of the event */ - type: "md_ip_adapter_list_output"; - }; - /** MDT2IAdapterListOutput */ - MDT2IAdapterListOutput: { + timestamp: number; /** - * T2I Adapter-List - * @description T2I-Adapter(s) to apply + * Id + * @description The ID of the install job */ - t2i_adapter_list: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + id: number; /** - * type - * @default md_ip_adapters_output - * @constant + * Source + * @description Source of the model; local path, repo_id or url */ - type: "md_ip_adapters_output"; - }; - /** - * MLSD Detection - * @description Generates an line segment map using MLSD. - */ - MLSDDetectionInvocation: { + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * @description The board to save the image to - * @default null + * Local Path + * @description Where model is downloading to */ - board?: components["schemas"]["BoardField"] | null; + local_path: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Bytes + * @description Number of bytes downloaded so far */ - metadata?: components["schemas"]["MetadataField"] | null; + bytes: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Total Bytes + * @description Total size of download, including all files */ - id: string; + total_bytes: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Parts + * @description Progress of downloading URLs that comprise the model, if any */ - is_intermediate?: boolean; + parts: { + [key: string]: number | string; + }[]; + }; + /** + * ModelInstallDownloadStartedEvent + * @description Event model for model_install_download_started + */ + ModelInstallDownloadStartedEvent: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Timestamp + * @description The timestamp of the event */ - use_cache?: boolean; + timestamp: number; /** - * @description The image to process - * @default null + * Id + * @description The ID of the install job */ - image?: components["schemas"]["ImageField"] | null; + id: number; /** - * Score Threshold - * @description The threshold used to score points when determining line segments - * @default 0.1 + * Source + * @description Source of the model; local path, repo_id or url */ - score_threshold?: number; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * Distance Threshold - * @description Threshold for including a line segment - lines shorter than this distance will be discarded - * @default 20 + * Local Path + * @description Where model is downloading to */ - distance_threshold?: number; + local_path: string; /** - * type - * @default mlsd_detection - * @constant + * Bytes + * @description Number of bytes downloaded so far */ - type: "mlsd_detection"; - }; - /** MainModelDefaultSettings */ - MainModelDefaultSettings: { + bytes: number; /** - * Vae - * @description Default VAE for this model (model key) + * Total Bytes + * @description Total size of download, including all files */ - vae?: string | null; + total_bytes: number; /** - * Vae Precision - * @description Default VAE precision for this model + * Parts + * @description Progress of downloading URLs that comprise the model, if any */ - vae_precision?: ("fp16" | "fp32") | null; + parts: { + [key: string]: number | string; + }[]; + }; + /** + * ModelInstallDownloadsCompleteEvent + * @description Emitted once when an install job becomes active. + */ + ModelInstallDownloadsCompleteEvent: { /** - * Scheduler - * @description Default scheduler for this model + * Timestamp + * @description The timestamp of the event */ - scheduler?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; + timestamp: number; /** - * Steps - * @description Default number of steps for this model + * Id + * @description The ID of the install job */ - steps?: number | null; + id: number; /** - * Cfg Scale - * @description Default CFG Scale for this model + * Source + * @description Source of the model; local path, repo_id or url */ - cfg_scale?: number | null; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + }; + /** + * ModelInstallErrorEvent + * @description Event model for model_install_error + */ + ModelInstallErrorEvent: { /** - * Cfg Rescale Multiplier - * @description Default CFG Rescale Multiplier for this model + * Timestamp + * @description The timestamp of the event */ - cfg_rescale_multiplier?: number | null; + timestamp: number; /** - * Width - * @description Default width for this model + * Id + * @description The ID of the install job */ - width?: number | null; + id: number; /** - * Height - * @description Default height for this model + * Source + * @description Source of the model; local path, repo_id or url */ - height?: number | null; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * Guidance - * @description Default Guidance for this model + * Error Type + * @description The name of the exception */ - guidance?: number | null; + error_type: string; /** - * Cpu Only - * @description Whether this model should run on CPU only + * Error + * @description A text description of the exception */ - cpu_only?: boolean | null; + error: string; }; /** - * Main Model - SD1.5, SD2 - * @description Loads a main model, outputting its submodels. + * ModelInstallJob + * @description Object that tracks the current status of an install request. */ - MainModelLoaderInvocation: { + ModelInstallJob: { /** * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description Unique ID for this job */ - id: string; + id: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * @description Current status of install process + * @default waiting */ - is_intermediate?: boolean; + status?: components["schemas"]["InstallStatus"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Error Reason + * @description Information about why the job failed */ - use_cache?: boolean; + error_reason?: string | null; + /** @description Configuration information (e.g. 'description') to apply to model. */ + config_in?: components["schemas"]["ModelRecordChanges"]; /** - * @description Main model (UNet, VAE, CLIP) to load - * @default null + * Config Out + * @description After successful installation, this will hold the configuration object. */ - model?: components["schemas"]["ModelIdentifierField"] | null; + config_out?: (components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]) | null; /** - * type - * @default main_model_loader - * @constant + * Inplace + * @description Leave model in its current location; otherwise install under models directory + * @default false */ - type: "main_model_loader"; - }; - /** - * Main_BnBNF4_FLUX_Config - * @description Model config for main checkpoint models. - */ - Main_BnBNF4_FLUX_Config: { + inplace?: boolean; /** - * Key - * @description A unique key for this model. + * Source + * @description Source (URL, repo_id, or local path) of model */ - key: string; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * Hash - * @description The hash of the model file(s). + * Local Path + * Format: path + * @description Path to locally-downloaded model; may be the same as the source */ - hash: string; + local_path: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Bytes + * @description For a remote model, the number of bytes downloaded so far (may not be available) + * @default 0 */ - path: string; + bytes?: number; /** - * File Size - * @description The size of the model in bytes. + * Total Bytes + * @description Total size of the model to be installed + * @default 0 */ - file_size: number; + total_bytes?: number; /** - * Name - * @description Name of the model. + * Source Metadata + * @description Metadata provided by the model source */ - name: string; + source_metadata?: (components["schemas"]["BaseMetadata"] | components["schemas"]["HuggingFaceMetadata"]) | null; /** - * Description - * @description Model description + * Download Parts + * @description Download jobs contributing to this install */ - description: string | null; + download_parts?: components["schemas"]["DownloadJob"][]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Error + * @description On an error condition, this field will contain the text of the exception */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + error?: string | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Error Traceback + * @description On an error condition, this field will contain the exception traceback */ - source_api_response: string | null; + error_traceback?: string | null; + }; + /** + * ModelInstallStartedEvent + * @description Event model for model_install_started + */ + ModelInstallStartedEvent: { /** - * Cover Image - * @description Url for image to preview model + * Timestamp + * @description The timestamp of the event */ - cover_image: string | null; + timestamp: number; /** - * Type - * @default main - * @constant + * Id + * @description The ID of the install job */ - type: "main"; + id: number; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Source + * @description Source of the model; local path, repo_id or url */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + }; + /** + * ModelLoadCompleteEvent + * @description Event model for model_load_complete + */ + ModelLoadCompleteEvent: { /** - * Config Path - * @description Path to the config for this model, if any. + * Timestamp + * @description The timestamp of the event */ - config_path: string | null; + timestamp: number; /** - * Base - * @default flux - * @constant + * Config + * @description The model's config */ - base: "flux"; + config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; /** - * Format - * @default bnb_quantized_nf4b - * @constant + * @description The submodel type, if any + * @default null */ - format: "bnb_quantized_nf4b"; - variant: components["schemas"]["FluxVariantType"]; + submodel_type: components["schemas"]["SubModelType"] | null; }; /** - * Main_Checkpoint_FLUX_Config - * @description Model config for main checkpoint models. + * ModelLoadStartedEvent + * @description Event model for model_load_started */ - Main_Checkpoint_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; + ModelLoadStartedEvent: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Timestamp + * @description The timestamp of the event */ - path: string; + timestamp: number; /** - * File Size - * @description The size of the model in bytes. + * Config + * @description The model's config */ - file_size: number; + config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; /** - * Name - * @description Name of the model. + * @description The submodel type, if any + * @default null */ - name: string; + submodel_type: components["schemas"]["SubModelType"] | null; + }; + /** + * ModelLoaderOutput + * @description Model loader output + */ + ModelLoaderOutput: { /** - * Description - * @description Model description + * VAE + * @description VAE */ - description: string | null; + vae: components["schemas"]["VAEField"]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default model_loader_output + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "model_loader_output"; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - source_api_response: string | null; + clip: components["schemas"]["CLIPField"]; /** - * Cover Image - * @description Url for image to preview model + * UNet + * @description UNet (scheduler, LoRAs) */ - cover_image: string | null; + unet: components["schemas"]["UNetField"]; + }; + /** + * Model Name Grabber + * @description Outputs the Model's name as a string. Use the Model Identifier node as input. + */ + ModelNameGrabberInvocation: { /** - * Type - * @default main - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - type: "main"; + id: string; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + is_intermediate?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - config_path: string | null; + use_cache?: boolean; /** - * Format - * @default checkpoint - * @constant + * Model + * @description Get this input from a Model Identifier node. + * @default null */ - format: "checkpoint"; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Base - * @default flux + * type + * @default model_name_grabber_invocation * @constant */ - base: "flux"; - variant: components["schemas"]["FluxVariantType"]; + type: "model_name_grabber_invocation"; }; /** - * Main_Checkpoint_Flux2_Config - * @description Model config for FLUX.2 checkpoint models (e.g. Klein). + * ModelNameGrabberOutput + * @description Model Name Grabber output */ - Main_Checkpoint_Flux2_Config: { + ModelNameGrabberOutput: { /** - * Key - * @description A unique key for this model. + * Name + * @description Name of the Model */ - key: string; + name: string; /** - * Hash - * @description The hash of the model file(s). + * type + * @default model_name_grabber_output + * @constant */ - hash: string; + type: "model_name_grabber_output"; + }; + /** + * Model Name to Model + * @description Converts a model identified by its string name to a Model + */ + ModelNameToModelInvocation: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * Model Name + * @description Exact model name (must match exactly one model) + * @default null */ - description: string | null; + model_name?: string | null; + /** + * type + * @default model_name_to_model + * @constant + */ + type: "model_name_to_model"; + }; + /** + * ModelRecordChanges + * @description A set of changes to apply to a model. + */ + ModelRecordChanges: { /** * Source - * @description The original source of the model (path, URL or repo_id). + * @description original source of the model */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + source?: string | null; + /** @description type of model source */ + source_type?: components["schemas"]["ModelSourceType"] | null; /** * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description metadata from remote source */ - source_api_response: string | null; + source_api_response?: string | null; /** - * Cover Image - * @description Url for image to preview model + * Name + * @description Name of the model. */ - cover_image: string | null; + name?: string | null; /** - * Type - * @default main - * @constant + * Path + * @description Path to the model. */ - type: "main"; + path?: string | null; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Description + * @description Model description */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + description?: string | null; + /** @description The base model. */ + base?: components["schemas"]["BaseModelType"] | null; + /** @description Type of model */ + type?: components["schemas"]["ModelType"] | null; /** - * Config Path - * @description Path to the config for this model, if any. + * Key + * @description Database ID for this model */ - config_path: string | null; + key?: string | null; /** - * Format - * @default checkpoint - * @constant + * Hash + * @description hash of model file */ - format: "checkpoint"; + hash?: string | null; /** - * Base - * @default flux2 - * @constant + * File Size + * @description Size of model file */ - base: "flux2"; - variant: components["schemas"]["Flux2VariantType"]; - }; - /** Main_Checkpoint_SD1_Config */ - Main_Checkpoint_SD1_Config: { + file_size?: number | null; /** - * Key - * @description A unique key for this model. + * Format + * @description format of model file */ - key: string; + format?: string | null; /** - * Hash - * @description The hash of the model file(s). + * Trigger Phrases + * @description Set of trigger phrases for this model */ - hash: string; + trigger_phrases?: string[] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Default Settings + * @description Default settings for this model */ - path: string; + default_settings?: components["schemas"]["MainModelDefaultSettings"] | components["schemas"]["LoraModelDefaultSettings"] | components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * File Size - * @description The size of the model in bytes. + * Cpu Only + * @description Whether this model should run on CPU only */ - file_size: number; + cpu_only?: boolean | null; /** - * Name - * @description Name of the model. + * Variant + * @description The variant of the model. */ - name: string; + variant?: components["schemas"]["ModelVariantType"] | components["schemas"]["ClipVariantType"] | components["schemas"]["FluxVariantType"] | components["schemas"]["Flux2VariantType"] | components["schemas"]["ZImageVariantType"] | components["schemas"]["Qwen3VariantType"] | null; + /** @description The prediction type of the model. */ + prediction_type?: components["schemas"]["SchedulerPredictionType"] | null; /** - * Description - * @description Model description + * Upcast Attention + * @description Whether to upcast attention. */ - description: string | null; + upcast_attention?: boolean | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Config Path + * @description Path to config file for model */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + config_path?: string | null; + }; + /** + * ModelRecordOrderBy + * @description The order in which to return model summaries. + * @enum {string} + */ + ModelRecordOrderBy: "default" | "type" | "base" | "name" | "format" | "size" | "created_at" | "updated_at" | "path"; + /** ModelRelationshipBatchRequest */ + ModelRelationshipBatchRequest: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Model Keys + * @description List of model keys to fetch related models for */ - source_api_response: string | null; + model_keys: string[]; + }; + /** ModelRelationshipCreateRequest */ + ModelRelationshipCreateRequest: { /** - * Cover Image - * @description Url for image to preview model + * Model Key 1 + * @description The key of the first model in the relationship */ - cover_image: string | null; + model_key_1: string; /** - * Type - * @default main - * @constant + * Model Key 2 + * @description The key of the second model in the relationship */ - type: "main"; + model_key_2: string; + }; + /** + * ModelRepoVariant + * @description Various hugging face variants on the diffusers format. + * @enum {string} + */ + ModelRepoVariant: "" | "fp16" | "fp32" | "onnx" | "openvino" | "flax"; + /** + * ModelSourceType + * @description Model source type. + * @enum {string} + */ + ModelSourceType: "path" | "url" | "hf_repo_id"; + /** + * Model To String + * @description Converts an Model to a JSONString + */ + ModelToStringInvocation: { /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + is_intermediate?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - config_path: string | null; + use_cache?: boolean; /** - * Format - * @default checkpoint - * @constant + * Model + * @description The model to select + * @default null */ - format: "checkpoint"; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Base - * @default sd-1 + * type + * @default model_to_string * @constant */ - base: "sd-1"; + type: "model_to_string"; }; - /** Main_Checkpoint_SD2_Config */ - Main_Checkpoint_SD2_Config: { + /** + * Model Toggle + * @description Allows boolean selection between two separate ModelIdentifier inputs + */ + ModelToggleInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Use Second + * @description Use 2nd Input + * @default null */ - file_size: number; + use_second?: boolean | null; /** - * Name - * @description Name of the model. + * @description First Model Input + * @default null */ - name: string; + model1?: components["schemas"]["ModelIdentifierField"] | null; /** - * Description - * @description Model description + * @description First Model Input + * @default null */ - description: string | null; + model2?: components["schemas"]["ModelIdentifierField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default model_toggle + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "model_toggle"; + }; + /** + * ModelToggleOutput + * @description Model Toggle output + */ + ModelToggleOutput: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Model + * @description Model identifier */ - source_api_response: string | null; + model: components["schemas"]["ModelIdentifierField"]; /** - * Cover Image - * @description Url for image to preview model + * type + * @default model_toggle_output + * @constant */ - cover_image: string | null; + type: "model_toggle_output"; + }; + /** + * ModelType + * @description Model type. + * @enum {string} + */ + ModelType: "onnx" | "main" | "vae" | "lora" | "control_lora" | "controlnet" | "embedding" | "ip_adapter" | "clip_vision" | "clip_embed" | "t2i_adapter" | "t5_encoder" | "qwen3_encoder" | "spandrel_image_to_image" | "siglip" | "flux_redux" | "llava_onevision" | "unknown"; + /** + * ModelVariantType + * @description Variant type. + * @enum {string} + */ + ModelVariantType: "normal" | "inpaint" | "depth"; + /** + * ModelsList + * @description Return list of configs. + */ + ModelsList: { + /** Models */ + models: (components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"])[]; + }; + /** + * MonochromeFilmGrain + * @description Adds monochrome film grain to an image + */ + MonochromeFilmGrainInvocation: { /** - * Type - * @default main - * @constant + * @description The board to save the image to + * @default null */ - type: "main"; + board?: components["schemas"]["BoardField"] | null; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * @description Optional metadata to be saved with the image + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Config Path - * @description Path to the config for this model, if any. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - config_path: string | null; + id: string; /** - * Format - * @default checkpoint - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - format: "checkpoint"; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; + is_intermediate?: boolean; /** - * Base - * @default sd-2 - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - base: "sd-2"; - }; - /** Main_Checkpoint_SDXLRefiner_Config */ - Main_Checkpoint_SDXLRefiner_Config: { + use_cache?: boolean; /** - * Key - * @description A unique key for this model. + * @description The image to add film grain to + * @default null */ - key: string; + image?: components["schemas"]["ImageField"] | null; /** - * Hash - * @description The hash of the model file(s). + * Amount 1 + * @description Amount of the first noise layer + * @default 100 */ - hash: string; + amount_1?: number; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Amount 2 + * @description Amount of the second noise layer + * @default 50 */ - path: string; + amount_2?: number; /** - * File Size - * @description The size of the model in bytes. + * Seed 1 + * @description The first seed to use (omit for random) + * @default null */ - file_size: number; + seed_1?: number | null; /** - * Name - * @description Name of the model. + * Seed 2 + * @description The second seed to use (omit for random) + * @default null */ - name: string; + seed_2?: number | null; /** - * Description - * @description Model description + * Blur 1 + * @description The strength of the first noise blur + * @default 0.5 */ - description: string | null; + blur_1?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Blur 2 + * @description The strength of the second noise blur + * @default 0.5 */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + blur_2?: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * type + * @default monochrome_film_grain + * @constant */ - source_api_response: string | null; + type: "monochrome_film_grain"; + }; + /** + * Multiply Integers + * @description Multiplies two numbers + */ + MultiplyInvocation: { /** - * Cover Image - * @description Url for image to preview model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - cover_image: string | null; + id: string; /** - * Type - * @default main - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - type: "main"; + is_intermediate?: boolean; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + use_cache?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * A + * @description The first number + * @default 0 */ - config_path: string | null; + a?: number; /** - * Format - * @default checkpoint - * @constant + * B + * @description The second number + * @default 0 */ - format: "checkpoint"; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; + b?: number; /** - * Base - * @default sdxl-refiner + * type + * @default mul * @constant */ - base: "sdxl-refiner"; + type: "mul"; }; - /** Main_Checkpoint_SDXL_Config */ - Main_Checkpoint_SDXL_Config: { + /** + * Nightmare Promptgen + * @description makes new friends + */ + NightmareInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default false */ - hash: string; + use_cache?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Prompt + * @description starting point for the generated prompt + * @default */ - path: string; + prompt?: string; /** - * File Size - * @description The size of the model in bytes. + * Max New Tokens + * @description the maximum allowed amount of new tokens to generate + * @default 300 */ - file_size: number; + max_new_tokens?: number; /** - * Name - * @description Name of the model. + * Min New Tokens + * @description the minimum new tokens - NOTE, this can increase generation time + * @default 30 */ - name: string; + min_new_tokens?: number; /** - * Description - * @description Model description + * Max Time + * @description Overrules min tokens; the max amount of time allowed to generate + * @default 10 */ - description: string | null; + max_time?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Temp + * @description Temperature + * @default 1.8 */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + temp?: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Typical P + * @description Lower than 1.0 seems crazier, higher is more consistent. + * @default 1 */ - source_api_response: string | null; + typical_p?: number; /** - * Cover Image - * @description Url for image to preview model + * Top P + * @description Top P sampling + * @default 0.9 */ - cover_image: string | null; + top_p?: number; /** - * Type - * @default main - * @constant + * Top K + * @description Top K sampling + * @default 20 */ - type: "main"; + top_k?: number; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Repetition Penalty + * @description Higher than 1.0 will try to prevent repetition. + * @default 1 */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + repetition_penalty?: number; /** - * Config Path - * @description Path to the config for this model, if any. + * Include Starter + * @description Include your prompt starter with the output or not. + * @default true */ - config_path: string | null; + include_starter?: boolean; /** - * Format - * @default checkpoint - * @constant + * Repo Id + * @default cactusfriend/nightmare-promptgen-3 + * @enum {string} */ - format: "checkpoint"; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; + repo_id?: "cactusfriend/nightmare-promptgen-3" | "cactusfriend/nightmare-promptgen-XL" | "cactusfriend/nightmare-invokeai-prompts"; /** - * Base - * @default sdxl + * type + * @default nightmare_promptgen * @constant */ - base: "sdxl"; + type: "nightmare_promptgen"; }; /** - * Main_Checkpoint_ZImage_Config - * @description Model config for Z-Image single-file checkpoint models (safetensors, etc). + * NightmareOutput + * @description Nightmare prompt string output */ - Main_Checkpoint_ZImage_Config: { + NightmareOutput: { /** - * Key - * @description A unique key for this model. + * Prompt + * @description The generated nightmare prompt string */ - key: string; + prompt: string; /** - * Hash - * @description The hash of the model file(s). + * type + * @default nightmare_str_output + * @constant */ - hash: string; + type: "nightmare_str_output"; + }; + /** NodeFieldValue */ + NodeFieldValue: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Node Path + * @description The node into which this batch data item will be substituted. */ - path: string; + node_path: string; /** - * File Size - * @description The size of the model in bytes. + * Field Name + * @description The field into which this batch data item will be substituted. */ - file_size: number; + field_name: string; /** - * Name - * @description Name of the model. + * Value + * @description The value to substitute into the node/field. */ - name: string; + value: string | number | components["schemas"]["ImageField"]; + }; + /** + * Add Noise (Flux) + * @description Add noise to a flux latents tensor using the appropriate ratio given the denoising schedule timestep. + * + * Calculates the correct initial timestep noising amount and applies it to the given latent tensor using simple addition according to the specified ratio. + */ + NoiseAddFluxInvocation: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * @description Latents tensor + * @default null */ - cover_image: string | null; + latents_in?: components["schemas"]["LatentsField"] | null; /** - * Type - * @default main - * @constant + * @description The noise to be added + * @default null */ - type: "main"; + noise_in?: components["schemas"]["LatentsField"] | null; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Num Steps + * @description Number of diffusion steps + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + num_steps?: number | null; /** - * Config Path - * @description Path to the config for this model, if any. + * Denoising Start + * @description Starting point for denoising (0.0 to 1.0) + * @default null */ - config_path: string | null; + denoising_start?: number | null; /** - * Base - * @default z-image - * @constant + * Is Schnell + * @description Boolean flag indicating if this is a FLUX Schnell model + * @default null */ - base: "z-image"; + is_schnell?: boolean | null; /** - * Format - * @default checkpoint + * type + * @default noise_add_flux * @constant */ - format: "checkpoint"; - variant: components["schemas"]["ZImageVariantType"]; + type: "noise_add_flux"; }; - /** Main_Diffusers_CogView4_Config */ - Main_Diffusers_CogView4_Config: { + /** + * 2D Noise Image + * @description Creates an image of 2D Noise approximating the desired characteristics. + * + * Creates an image of 2D Noise approximating the desired characteristics, using various combinations of gaussian blur and arithmetic operations to perform low pass and high pass filtering of 2-dimensional spatial frequencies of each channel to create Red, Blue, or Green "colored noise". + */ + NoiseImage2DInvocation: { /** - * Key - * @description A unique key for this model. + * @description The board to save the image to + * @default null */ - key: string; + board?: components["schemas"]["BoardField"] | null; /** - * Hash - * @description The hash of the model file(s). + * @description Optional metadata to be saved with the image + * @default null */ - hash: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * Noise Type + * @description Desired noise spectral characteristics + * @default White + * @enum {string} */ - description: string | null; + noise_type?: "White" | "Red" | "Blue" | "Green"; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Width + * @description Desired image width + * @default 512 */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + width?: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Height + * @description Desired image height + * @default 512 */ - source_api_response: string | null; + height?: number; /** - * Cover Image - * @description Url for image to preview model + * Seed + * @description Seed for noise generation + * @default 0 */ - cover_image: string | null; + seed?: number; /** - * Type - * @default main - * @constant + * Iterations + * @description Noise approx. iterations + * @default 15 */ - type: "main"; + iterations?: number; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Blur Threshold + * @description Threshold used in computing noise (lower is better/slower) + * @default 0.2 */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + blur_threshold?: number; /** - * Format - * @default diffusers - * @constant + * Sigma Red + * @description Sigma for strong gaussian blur LPF for red/green + * @default 3 */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + sigma_red?: number; /** - * Base - * @default cogview4 + * Sigma Blue + * @description Sigma for weak gaussian blur HPF for blue/green + * @default 1 + */ + sigma_blue?: number; + /** + * type + * @default noiseimg_2d * @constant */ - base: "cogview4"; + type: "noiseimg_2d"; }; /** - * Main_Diffusers_FLUX_Config - * @description Model config for FLUX.1 models in diffusers format. + * Create Latent Noise + * @description Generates latent noise. */ - Main_Diffusers_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + NoiseInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * Seed + * @description Seed for random number generation + * @default 0 */ - description: string | null; + seed?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Width + * @description Width of output (px) + * @default 512 */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + width?: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Height + * @description Height of output (px) + * @default 512 */ - source_api_response: string | null; + height?: number; /** - * Cover Image - * @description Url for image to preview model + * Use Cpu + * @description Use CPU for noise generation (for reproducible results across platforms) + * @default true */ - cover_image: string | null; + use_cpu?: boolean; /** - * Type - * @default main + * type + * @default noise * @constant */ - type: "main"; + type: "noise"; + }; + /** + * NoiseOutput + * @description Invocation noise output + */ + NoiseOutput: { + /** @description Noise tensor */ + noise: components["schemas"]["LatentsField"]; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Width + * @description Width of output (px) */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + width: number; /** - * Format - * @default diffusers - * @constant + * Height + * @description Height of output (px) */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + height: number; /** - * Base - * @default flux + * type + * @default noise_output * @constant */ - base: "flux"; - variant: components["schemas"]["FluxVariantType"]; + type: "noise_output"; }; /** - * Main_Diffusers_Flux2_Config - * @description Model config for FLUX.2 models in diffusers format (e.g. FLUX.2 Klein). + * Noise (Spectral characteristics) + * @description Creates a latents tensor of 2D noise channels approximating the desired characteristics. + * + * This operates like 2D Noise Image but outputs latent tensors, 4-channel or 16-channel. */ - Main_Diffusers_Flux2_Config: { + NoiseSpectralInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Noise Type + * @description Desired noise spectral characteristics + * @default White + * @enum {string} */ - file_size: number; + noise_type?: "White" | "Red" | "Blue" | "Green"; /** - * Name - * @description Name of the model. + * Width + * @description Desired image width + * @default 512 */ - name: string; + width?: number; /** - * Description - * @description Model description + * Height + * @description Desired image height + * @default 512 */ - description: string | null; + height?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Flux16 + * @description If false, 4-channel (SD/SDXL); if true, 16-channel (Flux) + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + flux16?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Seed + * @description Seed for noise generation + * @default 0 */ - source_api_response: string | null; + seed?: number; /** - * Cover Image - * @description Url for image to preview model + * Iterations + * @description Noise approx. iterations + * @default 15 */ - cover_image: string | null; + iterations?: number; /** - * Type - * @default main - * @constant + * Blur Threshold + * @description Threshold used in computing noise (lower is better/slower) + * @default 0.2 */ - type: "main"; + blur_threshold?: number; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Sigma Red + * @description Sigma for strong gaussian blur LPF for red/green + * @default 3 */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + sigma_red?: number; /** - * Format - * @default diffusers - * @constant + * Sigma Blue + * @description Sigma for weak gaussian blur HPF for blue/green + * @default 1 */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + sigma_blue?: number; /** - * Base - * @default flux2 + * type + * @default noise_spectral * @constant */ - base: "flux2"; - variant: components["schemas"]["Flux2VariantType"]; + type: "noise_spectral"; }; - /** Main_Diffusers_SD1_Config */ - Main_Diffusers_SD1_Config: { + /** + * Normal Map + * @description Generates a normal map. + */ + NormalMapInvocation: { /** - * Key - * @description A unique key for this model. + * @description The board to save the image to + * @default null */ - key: string; + board?: components["schemas"]["BoardField"] | null; /** - * Hash - * @description The hash of the model file(s). + * @description Optional metadata to be saved with the image + * @default null */ - hash: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * @description The image to process + * @default null */ - description: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default normal_map + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "normal_map"; + }; + /** + * Octree Quantizer + * @description Quantizes an image to the desired number of colors + */ + OctreeQuantizerInvocation: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description Optional metadata to be saved with the image + * @default null */ - source_api_response: string | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Cover Image - * @description Url for image to preview model + * @description The board to save the image to + * @default null */ - cover_image: string | null; + board?: components["schemas"]["BoardField"] | null; /** - * Type - * @default main - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - type: "main"; + id: string; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image to quantize + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Format - * @default diffusers - * @constant + * Final Colors + * @description The final number of colors in the palette + * @default 16 */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; + final_colors?: number; /** - * Base - * @default sd-1 + * type + * @default octree_quantizer * @constant */ - base: "sd-1"; + type: "octree_quantizer"; }; - /** Main_Diffusers_SD2_Config */ - Main_Diffusers_SD2_Config: { + /** + * Offset Latents + * @description Offsets a latents tensor by a given percentage of height/width. + * + * This takes a Latents input as well as two numbers (between 0 and 1), which are used to offset the latents in the vertical and/or horizontal directions. 0.5/0.5 would offset the image 50% in both directions such that the corners will wrap around and become the center of the image. + */ + OffsetLatentsInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * @description Latents tensor + * @default null */ - file_size: number; + latents?: components["schemas"]["LatentsField"] | null; /** - * Name - * @description Name of the model. + * X Offset + * @description Approx percentage to offset (H) + * @default 0.5 */ - name: string; + x_offset?: number; /** - * Description - * @description Model description + * Y Offset + * @description Approx percentage to offset (V) + * @default 0.5 */ - description: string | null; + y_offset?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default offset_latents + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "offset_latents"; + }; + /** OffsetPaginatedResults[BoardDTO] */ + OffsetPaginatedResults_BoardDTO_: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Limit + * @description Limit of items to get */ - source_api_response: string | null; + limit: number; /** - * Cover Image - * @description Url for image to preview model + * Offset + * @description Offset from which to retrieve items */ - cover_image: string | null; + offset: number; /** - * Type - * @default main - * @constant + * Total + * @description Total number of items in result */ - type: "main"; + total: number; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Items + * @description Items */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + items: components["schemas"]["BoardDTO"][]; + }; + /** OffsetPaginatedResults[ImageDTO] */ + OffsetPaginatedResults_ImageDTO_: { /** - * Format - * @default diffusers - * @constant + * Limit + * @description Limit of items to get */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; + limit: number; /** - * Base - * @default sd-2 - * @constant + * Offset + * @description Offset from which to retrieve items */ - base: "sd-2"; - }; - /** Main_Diffusers_SD3_Config */ - Main_Diffusers_SD3_Config: { + offset: number; /** - * Key - * @description A unique key for this model. + * Total + * @description Total number of items in result */ - key: string; + total: number; /** - * Hash - * @description The hash of the model file(s). + * Items + * @description Items */ - hash: string; + items: components["schemas"]["ImageDTO"][]; + }; + /** + * Optimized Tile Size From Area + * @description Cuts up an image into overlapping tiles, optimized for the constraints of min area, max area, overlap, and + * maximum tile dimension (either width or height). Returns ideal width and height. + */ + OptimizedTileSizeFromAreaInvocation: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * Width + * @description Image width in pixels + * @default null */ - description: string | null; + width?: number | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Height + * @description Image height in pixels + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + height?: number | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Min Area + * @description Minimum tile area + * @default 262144 */ - source_api_response: string | null; + min_area?: number; /** - * Cover Image - * @description Url for image to preview model + * Max Area + * @description Maximum tile area + * @default 1048576 */ - cover_image: string | null; + max_area?: number; /** - * Type - * @default main - * @constant + * Overlap + * @description Minimum overlap in pixels (both directions) + * @default null */ - type: "main"; + overlap?: number | null; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Max Tile Dim + * @description Maximum dimension of a tile (to limit artifacts) + * @default 2048 */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + max_tile_dim?: number; /** - * Format - * @default diffusers + * type + * @default optimized_tile_size_from_area * @constant */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + type: "optimized_tile_size_from_area"; + }; + /** + * OrphanedModelInfo + * @description Information about an orphaned model directory. + */ + OrphanedModelInfo: { /** - * Base - * @default sd-3 - * @constant + * Path + * @description Relative path to the orphaned directory from models root */ - base: "sd-3"; + path: string; /** - * Submodels - * @description Loadable submodels in this model + * Absolute Path + * @description Absolute path to the orphaned directory */ - submodels: { - [key: string]: components["schemas"]["SubmodelDefinition"]; - } | null; - }; - /** Main_Diffusers_SDXLRefiner_Config */ - Main_Diffusers_SDXLRefiner_Config: { + absolute_path: string; /** - * Key - * @description A unique key for this model. + * Files + * @description List of model files in this directory */ - key: string; + files: string[]; /** - * Hash - * @description The hash of the model file(s). + * Size Bytes + * @description Total size of all files in bytes */ - hash: string; + size_bytes: number; + }; + /** + * OutputFieldJSONSchemaExtra + * @description Extra attributes to be added to input fields and their OpenAPI schema. Used by the workflow editor + * during schema parsing and UI rendering. + */ + OutputFieldJSONSchemaExtra: { + field_kind: components["schemas"]["FieldKind"]; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Ui Hidden + * @default false */ - path: string; + ui_hidden: boolean; + /** + * Ui Order + * @default null + */ + ui_order: number | null; + /** @default null */ + ui_type: components["schemas"]["UIType"] | null; + }; + /** + * PBR Maps + * @description Generate Normal, Displacement and Roughness Map from a given image + */ + PBRMapsInvocation: { /** - * File Size - * @description The size of the model in bytes. + * @description The board to save the image to + * @default null */ - file_size: number; + board?: components["schemas"]["BoardField"] | null; /** - * Name - * @description Name of the model. + * @description Optional metadata to be saved with the image + * @default null */ - name: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * @description Input image + * @default null */ - cover_image: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Type - * @default main - * @constant + * Tile Size + * @description Tile size + * @default 512 */ - type: "main"; + tile_size?: number; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Border Mode + * @description Border mode to apply to eliminate any artifacts or seams + * @default none + * @enum {string} */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + border_mode?: "none" | "seamless" | "mirror" | "replicate"; /** - * Format - * @default diffusers + * type + * @default pbr_maps * @constant */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; + type: "pbr_maps"; + }; + /** PBRMapsOutput */ + PBRMapsOutput: { /** - * Base - * @default sdxl-refiner - * @constant + * @description The generated normal map + * @default null */ - base: "sdxl-refiner"; - }; - /** Main_Diffusers_SDXL_Config */ - Main_Diffusers_SDXL_Config: { + normal_map: components["schemas"]["ImageField"]; /** - * Key - * @description A unique key for this model. + * @description The generated roughness map + * @default null */ - key: string; + roughness_map: components["schemas"]["ImageField"]; /** - * Hash - * @description The hash of the model file(s). + * @description The generated displacement map + * @default null */ - hash: string; + displacement_map: components["schemas"]["ImageField"]; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * type + * @default pbr_maps-output + * @constant */ - path: string; + type: "pbr_maps-output"; + }; + /** + * PTFields Collect + * @description Collect Prompt Tools Fields for an image generated in InvokeAI. + */ + PTFieldsCollectInvocation: { /** - * File Size - * @description The size of the model in bytes. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - file_size: number; + id: string; /** - * Name - * @description Name of the model. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - name: string; + is_intermediate?: boolean; /** - * Description - * @description Model description + * Use Cache + * @description Whether or not to use the cache + * @default true */ - description: string | null; + use_cache?: boolean; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Positive Prompt + * @description The positive prompt parameter + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + positive_prompt?: string | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Positive Style Prompt + * @description The positive style prompt parameter + * @default null */ - source_api_response: string | null; + positive_style_prompt?: string | null; /** - * Cover Image - * @description Url for image to preview model + * Negative Prompt + * @description The negative prompt parameter + * @default null */ - cover_image: string | null; + negative_prompt?: string | null; /** - * Type - * @default main - * @constant + * Negative Style Prompt + * @description The negative prompt parameter + * @default null */ - type: "main"; + negative_style_prompt?: string | null; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Seed + * @description Seed for random number generation + * @default null */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + seed?: number | null; /** - * Format - * @default diffusers - * @constant + * Width + * @description Width of output (px) + * @default null */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; + width?: number | null; /** - * Base - * @default sdxl - * @constant + * Height + * @description Height of output (px) + * @default null */ - base: "sdxl"; - }; - /** - * Main_Diffusers_ZImage_Config - * @description Model config for Z-Image diffusers models (Z-Image-Turbo, Z-Image-Base). - */ - Main_Diffusers_ZImage_Config: { + height?: number | null; /** - * Key - * @description A unique key for this model. + * Steps + * @description Number of steps to run + * @default null */ - key: string; + steps?: number | null; /** - * Hash - * @description The hash of the model file(s). + * Cfg Scale + * @description Classifier-Free Guidance scale + * @default null */ - hash: string; + cfg_scale?: number | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default null */ - path: string; + denoising_start?: number | null; /** - * File Size - * @description The size of the model in bytes. + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default null */ - file_size: number; + denoising_end?: number | null; /** - * Name - * @description Name of the model. + * Scheduler + * @description Scheduler to use during inference + * @default null */ - name: string; + scheduler?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; /** - * Description - * @description Model description + * type + * @default pt_fields_collect + * @constant */ - description: string | null; + type: "pt_fields_collect"; + }; + /** + * PTFieldsCollectOutput + * @description PTFieldsCollect Output + */ + PTFieldsCollectOutput: { /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Pt Fields + * @description PTFields in Json Format */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + pt_fields: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * type + * @default pt_fields_collect_output + * @constant */ - source_api_response: string | null; + type: "pt_fields_collect_output"; + }; + /** + * PTFields Expand + * @description Save Expand PTFields into individual items + */ + PTFieldsExpandInvocation: { /** - * Cover Image - * @description Url for image to preview model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - cover_image: string | null; + id: string; /** - * Type - * @default main - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - type: "main"; + is_intermediate?: boolean; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + use_cache?: boolean; /** - * Format - * @default diffusers - * @constant + * Pt Fields + * @description PTFields in json Format + * @default null */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + pt_fields?: string | null; /** - * Base - * @default z-image + * type + * @default pt_fields_expand * @constant */ - base: "z-image"; - variant: components["schemas"]["ZImageVariantType"]; + type: "pt_fields_expand"; }; /** - * Main_GGUF_FLUX_Config - * @description Model config for main checkpoint models. + * PTFieldsExpandOutput + * @description Expand Prompt Tools Fields for an image generated in InvokeAI. */ - Main_GGUF_FLUX_Config: { + PTFieldsExpandOutput: { /** - * Key - * @description A unique key for this model. + * Positive Prompt + * @description The positive prompt */ - key: string; + positive_prompt: string; /** - * Hash - * @description The hash of the model file(s). + * Positive Style Prompt + * @description The positive style prompt */ - hash: string; + positive_style_prompt: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Negative Prompt + * @description The negative prompt */ - path: string; + negative_prompt: string; /** - * File Size - * @description The size of the model in bytes. + * Negative Style Prompt + * @description The negative prompt */ - file_size: number; + negative_style_prompt: string; /** - * Name - * @description Name of the model. + * Seed + * @description Seed for random number generation */ - name: string; + seed: number; /** - * Description - * @description Model description + * Width + * @description Width of output (px) */ - description: string | null; + width: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Height + * @description Height of output (px) */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + height: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Steps + * @description Number of steps to run */ - source_api_response: string | null; + steps: number; /** - * Cover Image - * @description Url for image to preview model + * Cfg Scale + * @description Classifier-Free Guidance scale */ - cover_image: string | null; + cfg_scale: number; /** - * Type - * @default main - * @constant + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps */ - type: "main"; + denoising_start: number; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + denoising_end: number; /** - * Config Path - * @description Path to the config for this model, if any. + * Scheduler + * @description Scheduler to use during inference + * @enum {string} */ - config_path: string | null; + scheduler: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** - * Base - * @default flux + * type + * @default pt_fields_expand_output * @constant */ - base: "flux"; + type: "pt_fields_expand_output"; + }; + /** PaginatedResults[WorkflowRecordListItemWithThumbnailDTO] */ + PaginatedResults_WorkflowRecordListItemWithThumbnailDTO_: { /** - * Format - * @default gguf_quantized - * @constant + * Page + * @description Current Page */ - format: "gguf_quantized"; - variant: components["schemas"]["FluxVariantType"]; + page: number; + /** + * Pages + * @description Total number of pages + */ + pages: number; + /** + * Per Page + * @description Number of items per page + */ + per_page: number; + /** + * Total + * @description Total number of items in result + */ + total: number; + /** + * Items + * @description Items + */ + items: components["schemas"]["WorkflowRecordListItemWithThumbnailDTO"][]; }; /** - * Main_GGUF_Flux2_Config - * @description Model config for GGUF-quantized FLUX.2 checkpoint models (e.g. Klein). + * Pair Tile with Image + * @description Pair an image with its tile properties. */ - Main_GGUF_Flux2_Config: { + PairTileImageInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * @description The tile image. + * @default null */ - file_size: number; + image?: components["schemas"]["ImageField"] | null; /** - * Name - * @description Name of the model. + * @description The tile properties. + * @default null */ - name: string; + tile?: components["schemas"]["Tile"] | null; /** - * Description - * @description Model description + * type + * @default pair_tile_image + * @constant */ - description: string | null; + type: "pair_tile_image"; + }; + /** PairTileImageOutput */ + PairTileImageOutput: { + /** @description A tile description with its corresponding image. */ + tile_with_image: components["schemas"]["TileWithImage"]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default pair_tile_image_output + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "pair_tile_image_output"; + }; + /** + * PaletteOutput + * @description Base class for Cell Fracture output + */ + PaletteOutput: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description The palette output + * @default null */ - source_api_response: string | null; + image?: components["schemas"]["ImageField"]; /** - * Cover Image - * @description Url for image to preview model + * type + * @default get_palette_output + * @constant */ - cover_image: string | null; + type: "get_palette_output"; + }; + /** + * Parse Weighted String + * @description Parses a string containing weighted terms (e.g. `(word)++` or `word-`) and returns the cleaned string, list of terms, their weights, and positions. + */ + ParseWeightedStringInvocation: { /** - * Type - * @default main - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - type: "main"; + id: string; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + is_intermediate?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - config_path: string | null; + use_cache?: boolean; /** - * Base - * @default flux2 - * @constant + * Text + * @description The input string containing weighted expressions + * @default null */ - base: "flux2"; + text?: string | null; /** - * Format - * @default gguf_quantized + * type + * @default parse_weighted_string * @constant */ - format: "gguf_quantized"; - variant: components["schemas"]["Flux2VariantType"]; + type: "parse_weighted_string"; }; /** - * Main_GGUF_ZImage_Config - * @description Model config for GGUF-quantized Z-Image transformer models. + * Paste Image into Bounding Box + * @description Paste the source image into the target image at the given bounding box. + * + * The source image must be the same size as the bounding box, and the bounding box must fit within the target image. */ - Main_GGUF_ZImage_Config: { + PasteImageIntoBoundingBoxInvocation: { /** - * Key - * @description A unique key for this model. + * @description The board to save the image to + * @default null */ - key: string; + board?: components["schemas"]["BoardField"] | null; /** - * Hash - * @description The hash of the model file(s). + * @description Optional metadata to be saved with the image + * @default null */ - hash: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * @description The image to paste + * @default null */ - description: string | null; + source_image?: components["schemas"]["ImageField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * @description The image to paste into + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + target_image?: components["schemas"]["ImageField"] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description The bounding box to paste the image into + * @default null */ - source_api_response: string | null; + bounding_box?: components["schemas"]["BoundingBoxField"] | null; /** - * Cover Image - * @description Url for image to preview model + * type + * @default paste_image_into_bounding_box + * @constant */ - cover_image: string | null; + type: "paste_image_into_bounding_box"; + }; + /** + * Percent To Float + * @description Converts a string to a float and divides it by 100. + */ + PercentToFloatInvocation: { /** - * Type - * @default main - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - type: "main"; + id: string; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + is_intermediate?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - config_path: string | null; + use_cache?: boolean; /** - * Base - * @default z-image - * @constant + * Text + * @description Input text + * @default null */ - base: "z-image"; + text?: string | null; /** - * Format - * @default gguf_quantized + * type + * @default percent_to_float * @constant */ - format: "gguf_quantized"; - variant: components["schemas"]["ZImageVariantType"]; + type: "percent_to_float"; }; /** - * Combine Masks - * @description Combine two masks together by multiplying them using `PIL.ImageChops.multiply()`. + * PiDiNet Edge Detection + * @description Generates an edge map using PiDiNet. */ - MaskCombineInvocation: { + PiDiNetEdgeDetectionInvocation: { /** * @description The board to save the image to * @default null @@ -19194,27 +32925,34 @@ export type components = { */ use_cache?: boolean; /** - * @description The first mask to combine + * @description The image to process * @default null */ - mask1?: components["schemas"]["ImageField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description The second image to combine - * @default null + * Quantize Edges + * @description Whether or not to use safe mode + * @default false */ - mask2?: components["schemas"]["ImageField"] | null; + quantize_edges?: boolean; + /** + * Scribble + * @description Whether or not to use scribble mode + * @default false + */ + scribble?: boolean; /** * type - * @default mask_combine + * @default pidi_edge_detection * @constant */ - type: "mask_combine"; + type: "pidi_edge_detection"; }; /** - * Mask Edge - * @description Applies an edge mask to an image + * Pixel Art + * @description Convert an image to pixelart */ - MaskEdgeInvocation: { + PixelArtInvocation: { /** * @description The board to save the image to * @default null @@ -19243,46 +32981,45 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to apply the mask to + * @description The image to convert to pixelart * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Edge Size - * @description The size of the edge + * @description Optional - Image with the palette to use * @default null */ - edge_size?: number | null; + palette_image?: components["schemas"]["ImageField"] | null; /** - * Edge Blur - * @description The amount of blur on the edge - * @default null + * Pixel Size + * @description Size of the large 'pixels' in the output + * @default 4 */ - edge_blur?: number | null; + pixel_size?: number; /** - * Low Threshold - * @description First threshold for the hysteresis procedure in Canny edge detection - * @default null + * Palette Size + * @description No. of colours in Palette (used if palette_image is not provided or to extract from it) + * @default 16 */ - low_threshold?: number | null; + palette_size?: number; /** - * High Threshold - * @description Second threshold for the hysteresis procedure in Canny edge detection - * @default null + * Dither + * @description Apply Floyd-Steinberg dithering + * @default false */ - high_threshold?: number | null; + dither?: boolean; /** * type - * @default mask_edge + * @default pixel-art * @constant */ - type: "mask_edge"; + type: "pixel-art"; }; /** - * Mask from Alpha - * @description Extracts the alpha channel of an image as a mask. + * Pixelize Image + * @description Creates 'pixel' 'art' using trained models */ - MaskFromAlphaInvocation: { + PixelizeImageInvocation: { /** * @description The board to save the image to * @default null @@ -19311,33 +33048,34 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to create the mask from + * @description The image to pixelize * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Invert - * @description Whether or not to invert the mask - * @default false + * Correct Colors + * @description Correct hue and saturation to match original image. + * @default true */ - invert?: boolean; + correct_colors?: boolean; + /** + * Cell Size + * @description pixel/cell size (min 2 max WHATEVER BRO) + * @default 4 + */ + cell_size?: number; /** * type - * @default tomask + * @default pixelize * @constant */ - type: "tomask"; + type: "pixelize"; }; /** - * Mask from Segmented Image - * @description Generate a mask for a particular color in an ID Map + * Pixelize + * @description Pixelize an image. Downsample, upsample. */ - MaskFromIDInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + PixelizeInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -19361,73 +33099,132 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to create the mask from + * @description Input image for pixelization * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description ID color to mask - * @default null + * Downsample Factor + * @description Image resizing factor. Higher = smaller image. + * @default 4 */ - color?: components["schemas"]["ColorField"] | null; + downsample_factor?: number; /** - * Threshold - * @description Threshold for color detection - * @default 100 + * Upsample + * @description Upsample to original resolution + * @default true */ - threshold?: number; + upsample?: boolean; /** - * Invert - * @description Whether or not to invert the mask + * type + * @default retro_pixelize + * @constant + */ + type: "retro_pixelize"; + }; + /** PresetData */ + PresetData: { + /** + * Positive Prompt + * @description Positive prompt + */ + positive_prompt: string; + /** + * Negative Prompt + * @description Negative prompt + */ + negative_prompt: string; + }; + /** + * PresetType + * @enum {string} + */ + PresetType: "user" | "default"; + /** + * Print String to Console + * @description Prints a string to the console. + */ + PrintStringToConsoleInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. * @default false */ - invert?: boolean; + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Print Colour + * @description The colour to print the console output + * @default White on Black + * @enum {string} + */ + print_colour?: "Black on White" | "White on Black" | "Black on Green" | "Red on White" | "Green on Red" | "Yellow on Blue" | "Blue on Yellow" | "Magenta on Cyan" | "Cyan on Magenta" | "Black on Red" | "Red on Black" | "Yellow on Black" | "Black on Yellow"; + /** + * Input string + * @description The string to print to console. + * @default , + */ + input_str?: string; /** * type - * @default mask_from_id + * @default print_string_to_console_invocation * @constant */ - type: "mask_from_id"; + type: "print_string_to_console_invocation"; }; /** - * MaskOutput - * @description A torch mask tensor. + * PrintStringToConsoleOutput + * @description Debug Print String Output */ - MaskOutput: { - /** @description The mask. */ - mask: components["schemas"]["TensorField"]; - /** - * Width - * @description The width of the mask in pixels. - */ - width: number; + PrintStringToConsoleOutput: { /** - * Height - * @description The height of the mask in pixels. + * Passthrough + * @description Passthrough */ - height: number; + passthrough: string; /** * type - * @default mask_output + * @default print_string_to_console_output * @constant */ - type: "mask_output"; + type: "print_string_to_console_output"; }; /** - * Tensor Mask to Image - * @description Convert a mask tensor to an image. + * ProgressImage + * @description The progress image sent intermittently during processing */ - MaskTensorToImageInvocation: { + ProgressImage: { /** - * @description The board to save the image to - * @default null + * Width + * @description The effective width of the image in pixels */ - board?: components["schemas"]["BoardField"] | null; + width: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Height + * @description The effective height of the image in pixels */ - metadata?: components["schemas"]["MetadataField"] | null; + height: number; + /** + * Dataurl + * @description The image data as a b64 data URL + */ + dataURL: string; + }; + /** + * Prompt Auto .and() + * @description Takes a prompt string then chunks it up into a .and() output if over the max length + */ + PromptAutoAndInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -19446,32 +33243,29 @@ export type components = { */ use_cache?: boolean; /** - * @description The mask tensor to convert. - * @default null + * Prompt + * @description Prompt to auto .and() + * @default */ - mask?: components["schemas"]["TensorField"] | null; + prompt?: string; + /** + * Max Length + * @description Maximum chunk length in characters + * @default 200 + */ + max_length?: number; /** * type - * @default tensor_mask_to_image + * @default prompt_auto_and * @constant */ - type: "tensor_mask_to_image"; + type: "prompt_auto_and"; }; /** - * MediaPipe Face Detection - * @description Detects faces using MediaPipe. + * Prompt from Lookup Table + * @description Creates prompts using lookup table templates */ - MediaPipeFaceDetectionInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + PromptFromLookupTableInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -19486,38 +33280,59 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * @description The image to process - * @default null + * Resolutions Dict + * @description Private field for id substitutions dict cache + * @default {} */ - image?: components["schemas"]["ImageField"] | null; + resolutions_dict?: { + [key: string]: unknown; + }; /** - * Max Faces - * @description Maximum number of faces to detect - * @default 1 + * Lookups + * @description Lookup table(s) containing template(s) (JSON) + * @default [] */ - max_faces?: number; + lookups?: string | string[]; /** - * Min Confidence - * @description Minimum confidence for face detection - * @default 0.5 + * Remove Negatives + * @description Whether to strip out text between [] + * @default false */ - min_confidence?: number; + remove_negatives?: boolean; + /** + * Strip Parens Probability + * @description Probability of removing attention group weightings + * @default 0 + */ + strip_parens_probability?: number; + /** + * Resolutions + * @description JSON structure of substitutions by id by tag + * @default [] + */ + resolutions?: string | string[]; + /** + * Seed Vector In + * @description Optional JSON array of seeds for deterministic generation + * @default null + */ + seed_vector_in?: string | null; /** * type - * @default mediapipe_face_detection + * @default prompt_from_lookup_table * @constant */ - type: "mediapipe_face_detection"; + type: "prompt_from_lookup_table"; }; /** - * Metadata Merge - * @description Merged a collection of MetadataDict into a single MetadataDict. + * Prompt Strength + * @description Takes a prompt string and float strength and outputs a new string in the format of (prompt)strength */ - MergeMetadataInvocation: { + PromptStrengthInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -19537,32 +33352,56 @@ export type components = { use_cache?: boolean; /** * Collection - * @description Collection of Metadata - * @default null + * @description Collection of Prompt strengths + * @default [] */ - collection?: components["schemas"]["MetadataField"][] | null; + collection?: string[]; + /** + * Prompt + * @description Prompt to work on + * @default + */ + prompt?: string; + /** + * Strength + * @description strength of the prompt + * @default 1 + */ + strength?: number; /** * type - * @default merge_metadata + * @default prompt_strength * @constant */ - type: "merge_metadata"; + type: "prompt_strength"; }; /** - * Merge Tiles to Image - * @description Merge multiple tile images into a single image. + * PromptStrengthOutput + * @description Base class for nodes that output a collection of images */ - MergeTilesToImageInvocation: { + PromptStrengthOutput: { /** - * @description The board to save the image to - * @default null + * Collection + * @description Prompt strength collection */ - board?: components["schemas"]["BoardField"] | null; + collection: string[]; /** - * @description Optional metadata to be saved with the image - * @default null + * Value + * @description Prompt strength */ - metadata?: components["schemas"]["MetadataField"] | null; + value: string; + /** + * type + * @default prompt_strength_collection_output + * @constant + */ + type: "prompt_strength_collection_output"; + }; + /** + * Prompt Strengths Combine + * @description Takes a collection of prompt strength strings and converts it into a combined .and() or .blend() structure. Blank prompts are ignored + */ + PromptStrengthsCombineInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -19581,43 +33420,35 @@ export type components = { */ use_cache?: boolean; /** - * Tiles With Images - * @description A list of tile images with tile properties. - * @default null + * Prompt Strengths + * @description Prompt strengths to combine + * @default [ + * "" + * ] */ - tiles_with_images?: components["schemas"]["TileWithImage"][] | null; + prompt_strengths?: string[]; /** - * Blend Mode - * @description blending type Linear or Seam - * @default Seam + * Combine Type + * @description Combine type .and() or .blend() + * @default .and * @enum {string} */ - blend_mode?: "Linear" | "Seam"; - /** - * Blend Amount - * @description The amount to blend adjacent tiles in pixels. Must be <= the amount of overlap between adjacent tiles. - * @default 32 - */ - blend_amount?: number; + combine_type?: ".and" | ".blend"; /** * type - * @default merge_tiles_to_image + * @default prompt_strengths_combine * @constant */ - type: "merge_tiles_to_image"; + type: "prompt_strengths_combine"; }; /** - * MetadataField - * @description Pydantic model for metadata with custom root of type dict[str, Any]. - * Metadata is stored without a strict schema. - */ - MetadataField: Record; - /** - * Metadata Field Extractor - * @description Extracts the text value from an image's metadata given a key. - * Raises an error if the image has no metadata or if the value is not a string (nesting not permitted). + * Prompt Template + * @description Applies a Style Preset template to positive and negative prompts. + * + * Select a Style Preset and provide positive/negative prompts. The node replaces + * {prompt} placeholders in the template with your input prompts. */ - MetadataFieldExtractorInvocation: { + PromptTemplateInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -19636,62 +33467,56 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to extract metadata from + * @description The Style Preset to use as a template * @default null */ - image?: components["schemas"]["ImageField"] | null; + style_preset?: components["schemas"]["StylePresetField"] | null; /** - * Key - * @description The key in the image's metadata to extract the value from - * @default null + * Positive Prompt + * @description The positive prompt to insert into the template's {prompt} placeholder + * @default */ - key?: string | null; + positive_prompt?: string; + /** + * Negative Prompt + * @description The negative prompt to insert into the template's {prompt} placeholder + * @default + */ + negative_prompt?: string; /** * type - * @default metadata_field_extractor + * @default prompt_template * @constant */ - type: "metadata_field_extractor"; + type: "prompt_template"; }; /** - * Metadata From Image - * @description Used to create a core metadata item then Add/Update it to the provided metadata + * PromptTemplateOutput + * @description Output for the Prompt Template node */ - MetadataFromImageInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + PromptTemplateOutput: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Positive Prompt + * @description The positive prompt with the template applied */ - use_cache?: boolean; + positive_prompt: string; /** - * @description The image to process - * @default null + * Negative Prompt + * @description The negative prompt with the template applied */ - image?: components["schemas"]["ImageField"] | null; + negative_prompt: string; /** * type - * @default metadata_from_image + * @default prompt_template_output * @constant */ - type: "metadata_from_image"; + type: "prompt_template_output"; }; /** - * Metadata - * @description Takes a MetadataItem or collection of MetadataItems and outputs a MetadataDict. + * Prompts from File + * @description Loads prompts from a text file */ - MetadataInvocation: { + PromptsFromFileInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -19710,36 +33535,47 @@ export type components = { */ use_cache?: boolean; /** - * Items - * @description A single metadata item or collection of metadata items + * File Path + * @description Path to prompt text file * @default null */ - items?: components["schemas"]["MetadataItemField"][] | components["schemas"]["MetadataItemField"] | null; + file_path?: string | null; /** - * type - * @default metadata - * @constant + * Pre Prompt + * @description String to prepend to each prompt + * @default null */ - type: "metadata"; - }; - /** MetadataItemField */ - MetadataItemField: { + pre_prompt?: string | null; /** - * Label - * @description Label for this metadata item + * Post Prompt + * @description String to append to each prompt + * @default null */ - label: string; + post_prompt?: string | null; /** - * Value - * @description The value for this metadata item (may be any type) + * Start Line + * @description Line in the file to start start from + * @default 1 */ - value: unknown; + start_line?: number; + /** + * Max Prompts + * @description Max lines to read from file (0=all) + * @default 1 + */ + max_prompts?: number; + /** + * type + * @default prompt_from_file + * @constant + */ + type: "prompt_from_file"; }; /** - * Metadata Item - * @description Used to create an arbitrary metadata item. Provide "label" and make a connection to "value" to store that data as the value. + * Prompts To File + * @description Save prompts to a text file */ - MetadataItemInvocation: { + PromptsToFileInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -19758,34 +33594,58 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item + * File Path + * @description Path to prompt text file * @default null */ - label?: string | null; + file_path?: string | null; /** - * Value - * @description The value for this metadata item (may be any type) + * Prompts + * @description Prompt or collection of prompts to write * @default null */ - value?: unknown | null; + prompts?: string | string[] | null; + /** + * Append + * @description Append or overwrite file + * @default true + */ + append?: boolean; + /** + * type + * @default prompt_to_file + * @constant + */ + type: "prompt_to_file"; + }; + /** + * PromptsToFileInvocationOutput + * @description Base class for invocation that writes to a file and returns nothing of use + */ + PromptsToFileInvocationOutput: { /** * type - * @default metadata_item + * @default prompt_to_file_output * @constant */ - type: "metadata_item"; + type: "prompt_to_file_output"; }; /** - * Metadata Item Linked - * @description Used to Create/Add/Update a value into a metadata label + * PruneResult + * @description Result of pruning the session queue */ - MetadataItemLinkedInvocation: { + PruneResult: { /** - * @description Optional metadata to be saved with the image - * @default null + * Deleted + * @description Number of queue items deleted */ - metadata?: components["schemas"]["MetadataField"] | null; + deleted: number; + }; + /** + * Prune/Clean Prompts + * @description Like home staging but for prompt text + */ + PruneTextInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -19804,410 +33664,483 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Content + * @description Text to prune/cleanup + * @default null */ - label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt" | "width" | "height" | "seed" | "cfg_scale" | "cfg_rescale_multiplier" | "steps" | "scheduler" | "clip_skip" | "model" | "vae" | "seamless_x" | "seamless_y" | "guidance" | "cfg_scale_start_step" | "cfg_scale_end_step"; + content?: string | null; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Blacklist File Path + * @description Path to .txt to prune with. No path will run without matched content removal. + * @default */ - custom_label?: string | null; + blacklist_file_path?: string | null; /** - * Value - * @description The value for this metadata item (may be any type) - * @default null + * Remove Weight Syntax + * @description Remove basic Compel + A111-style attention weighting syntax. Special Compel syntax like .and() not supported + * @default false */ - value?: unknown | null; + remove_weight_syntax?: boolean; + /** + * Dedupe Tags + * @description Group text by commas, remove duplicates + * @default true + */ + dedupe_tags?: boolean; + /** + * Remove Slashes + * @description Delete all backslashes instead of just extras + * @default true + */ + remove_slashes?: boolean; + /** + * Remove Tis And Loras + * @description Delete Invoke TI syntax, A111 LoRAs, and Invoke 2.x LoRAs + * @default false + */ + remove_tis_and_loras?: boolean; + /** + * Custom Regex Pattern + * @description Custom regex pattern to apply to the text + * @default + */ + custom_regex_pattern?: string | null; + /** + * Custom Regex Substitution + * @description Substitution string for the custom regex pattern. Leave blank to simply delete captured text + * @default + */ + custom_regex_substitution?: string | null; /** * type - * @default metadata_item_linked + * @default prune_prompt * @constant */ - type: "metadata_item_linked"; + type: "prune_prompt"; }; /** - * MetadataItemOutput - * @description Metadata Item Output + * PrunedPromptOutput + * @description Pruned and cleaned string output */ - MetadataItemOutput: { - /** @description Metadata Item */ - item: components["schemas"]["MetadataItemField"]; + PrunedPromptOutput: { /** - * type - * @default metadata_item_output - * @constant + * Prompt + * @description Processed prompt string */ - type: "metadata_item_output"; - }; - /** MetadataOutput */ - MetadataOutput: { - /** @description Metadata Dict */ - metadata: components["schemas"]["MetadataField"]; + prompt: string; /** * type - * @default metadata_output + * @default clean_prompt * @constant */ - type: "metadata_output"; + type: "clean_prompt"; }; /** - * Metadata To Bool Collection - * @description Extracts a Boolean value Collection of a label from metadata + * QueueClearedEvent + * @description Event model for queue_cleared */ - MetadataToBoolCollectionInvocation: { + QueueClearedEvent: { /** - * @description Optional metadata to be saved with the image - * @default null + * Timestamp + * @description The timestamp of the event */ - metadata?: components["schemas"]["MetadataField"] | null; + timestamp: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Queue Id + * @description The ID of the queue */ - id: string; + queue_id: string; + }; + /** + * QueueItemStatusChangedEvent + * @description Event model for queue_item_status_changed + */ + QueueItemStatusChangedEvent: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Timestamp + * @description The timestamp of the event */ - is_intermediate?: boolean; + timestamp: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Queue Id + * @description The ID of the queue */ - use_cache?: boolean; + queue_id: string; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Item Id + * @description The ID of the queue item */ - label?: "* CUSTOM LABEL *" | "seamless_x" | "seamless_y"; + item_id: number; /** - * Custom Label - * @description Label for this metadata item + * Batch Id + * @description The ID of the queue batch + */ + batch_id: string; + /** + * Origin + * @description The origin of the queue item * @default null */ - custom_label?: string | null; + origin: string | null; /** - * Default Value - * @description The default bool to use if not found in the metadata + * Destination + * @description The destination of the queue item * @default null */ - default_value?: boolean[] | null; + destination: string | null; /** - * type - * @default metadata_to_bool_collection - * @constant + * User Id + * @description The ID of the user who created the queue item + * @default system */ - type: "metadata_to_bool_collection"; - }; - /** - * Metadata To Bool - * @description Extracts a Boolean value of a label from metadata - */ - MetadataToBoolInvocation: { + user_id: string; /** - * @description Optional metadata to be saved with the image + * Status + * @description The new status of the queue item + * @enum {string} + */ + status: "pending" | "in_progress" | "completed" | "failed" | "canceled"; + /** + * Error Type + * @description The error type, if any * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + error_type: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Error Message + * @description The error message, if any + * @default null */ - id: string; + error_message: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Error Traceback + * @description The error traceback, if any + * @default null */ - is_intermediate?: boolean; + error_traceback: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Created At + * @description The timestamp when the queue item was created */ - use_cache?: boolean; + created_at: string; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Updated At + * @description The timestamp when the queue item was last updated */ - label?: "* CUSTOM LABEL *" | "seamless_x" | "seamless_y"; + updated_at: string; /** - * Custom Label - * @description Label for this metadata item + * Started At + * @description The timestamp when the queue item was started * @default null */ - custom_label?: string | null; + started_at: string | null; /** - * Default Value - * @description The default bool to use if not found in the metadata + * Completed At + * @description The timestamp when the queue item was completed * @default null */ - default_value?: boolean | null; + completed_at: string | null; + /** @description The status of the batch */ + batch_status: components["schemas"]["BatchStatus"]; + /** @description The status of the queue */ + queue_status: components["schemas"]["SessionQueueStatus"]; /** - * type - * @default metadata_to_bool - * @constant + * Session Id + * @description The ID of the session (aka graph execution state) */ - type: "metadata_to_bool"; + session_id: string; }; /** - * Metadata To ControlNets - * @description Extracts a Controlnets value of a label from metadata + * QueueItemsRetriedEvent + * @description Event model for queue_items_retried */ - MetadataToControlnetsInvocation: { + QueueItemsRetriedEvent: { /** - * @description Optional metadata to be saved with the image - * @default null + * Timestamp + * @description The timestamp of the event */ - metadata?: components["schemas"]["MetadataField"] | null; + timestamp: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Queue Id + * @description The ID of the queue + */ + queue_id: string; + /** + * Retried Item Ids + * @description The IDs of the queue items that were retried + */ + retried_item_ids: number[]; + }; + /** + * Qwen3EncoderField + * @description Field for Qwen3 text encoder used by Z-Image models. + */ + Qwen3EncoderField: { + /** @description Info to load tokenizer submodel */ + tokenizer: components["schemas"]["ModelIdentifierField"]; + /** @description Info to load text_encoder submodel */ + text_encoder: components["schemas"]["ModelIdentifierField"]; + /** + * Loras + * @description LoRAs to apply on model loading + */ + loras?: components["schemas"]["LoRAField"][]; + }; + /** + * Qwen3Encoder_Checkpoint_Config + * @description Configuration for single-file Qwen3 Encoder models (safetensors). + */ + Qwen3Encoder_Checkpoint_Config: { + /** + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * ControlNet-List - * @default null + * File Size + * @description The size of the model in bytes. */ - control_list?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + file_size: number; /** - * type - * @default metadata_to_controlnets - * @constant + * Name + * @description Name of the model. */ - type: "metadata_to_controlnets"; - }; - /** - * Metadata To Float Collection - * @description Extracts a Float value Collection of a label from metadata - */ - MetadataToFloatCollectionInvocation: { + name: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Description + * @description Model description */ - metadata?: components["schemas"]["MetadataField"] | null; + description: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source + * @description The original source of the model (path, URL or repo_id). */ - id: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - is_intermediate?: boolean; + source_api_response: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Cover Image + * @description Url for image to preview model */ - use_cache?: boolean; + cover_image: string | null; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Config Path + * @description Path to the config for this model, if any. */ - label?: "* CUSTOM LABEL *" | "cfg_scale" | "cfg_rescale_multiplier" | "guidance"; + config_path: string | null; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Base + * @default any + * @constant */ - custom_label?: string | null; + base: "any"; /** - * Default Value - * @description The default float to use if not found in the metadata - * @default null + * Type + * @default qwen3_encoder + * @constant */ - default_value?: number[] | null; + type: "qwen3_encoder"; /** - * type - * @default metadata_to_float_collection + * Format + * @default checkpoint * @constant */ - type: "metadata_to_float_collection"; + format: "checkpoint"; + /** + * Cpu Only + * @description Whether this model should run on CPU only + */ + cpu_only: boolean | null; + /** @description Qwen3 model size variant (4B or 8B) */ + variant: components["schemas"]["Qwen3VariantType"]; }; /** - * Metadata To Float - * @description Extracts a Float value of a label from metadata + * Qwen3Encoder_GGUF_Config + * @description Configuration for GGUF-quantized Qwen3 Encoder models. */ - MetadataToFloatInvocation: { + Qwen3Encoder_GGUF_Config: { /** - * @description Optional metadata to be saved with the image - * @default null + * Key + * @description A unique key for this model. */ - metadata?: components["schemas"]["MetadataField"] | null; + key: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Hash + * @description The hash of the model file(s). */ - id: string; + hash: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - is_intermediate?: boolean; + path: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * File Size + * @description The size of the model in bytes. */ - use_cache?: boolean; + file_size: number; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Name + * @description Name of the model. */ - label?: "* CUSTOM LABEL *" | "cfg_scale" | "cfg_rescale_multiplier" | "guidance"; + name: string; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Description + * @description Model description */ - custom_label?: string | null; + description: string | null; /** - * Default Value - * @description The default float to use if not found in the metadata - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - default_value?: number | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * type - * @default metadata_to_float - * @constant + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - type: "metadata_to_float"; - }; - /** - * Metadata To IP-Adapters - * @description Extracts a IP-Adapters value of a label from metadata - */ - MetadataToIPAdaptersInvocation: { + source_api_response: string | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Cover Image + * @description Url for image to preview model */ - metadata?: components["schemas"]["MetadataField"] | null; + cover_image: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Config Path + * @description Path to the config for this model, if any. */ - id: string; + config_path: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Base + * @default any + * @constant */ - is_intermediate?: boolean; + base: "any"; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Type + * @default qwen3_encoder + * @constant */ - use_cache?: boolean; + type: "qwen3_encoder"; /** - * IP-Adapter-List - * @description IP-Adapter to apply - * @default null + * Format + * @default gguf_quantized + * @constant */ - ip_adapter_list?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + format: "gguf_quantized"; /** - * type - * @default metadata_to_ip_adapters - * @constant + * Cpu Only + * @description Whether this model should run on CPU only */ - type: "metadata_to_ip_adapters"; + cpu_only: boolean | null; + /** @description Qwen3 model size variant (4B or 8B) */ + variant: components["schemas"]["Qwen3VariantType"]; }; /** - * Metadata To Integer Collection - * @description Extracts an integer value Collection of a label from metadata + * Qwen3Encoder_Qwen3Encoder_Config + * @description Configuration for Qwen3 Encoder models in a diffusers-like format. + * + * The model weights are expected to be in a folder called text_encoder inside the model directory, + * compatible with Qwen2VLForConditionalGeneration or similar architectures used by Z-Image. */ - MetadataToIntegerCollectionInvocation: { + Qwen3Encoder_Qwen3Encoder_Config: { /** - * @description Optional metadata to be saved with the image - * @default null + * Key + * @description A unique key for this model. */ - metadata?: components["schemas"]["MetadataField"] | null; + key: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Hash + * @description The hash of the model file(s). */ - id: string; + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - is_intermediate?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - use_cache?: boolean; + source_api_response: string | null; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Cover Image + * @description Url for image to preview model */ - label?: "* CUSTOM LABEL *" | "width" | "height" | "seed" | "steps" | "clip_skip" | "cfg_scale_start_step" | "cfg_scale_end_step"; + cover_image: string | null; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Base + * @default any + * @constant */ - custom_label?: string | null; + base: "any"; /** - * Default Value - * @description The default integer to use if not found in the metadata - * @default null + * Type + * @default qwen3_encoder + * @constant */ - default_value?: number[] | null; + type: "qwen3_encoder"; /** - * type - * @default metadata_to_integer_collection + * Format + * @default qwen3_encoder * @constant */ - type: "metadata_to_integer_collection"; + format: "qwen3_encoder"; + /** + * Cpu Only + * @description Whether this model should run on CPU only + */ + cpu_only: boolean | null; + /** @description Qwen3 model size variant (4B or 8B) */ + variant: components["schemas"]["Qwen3VariantType"]; }; /** - * Metadata To Integer - * @description Extracts an integer value of a label from metadata + * Qwen3 Prompt Pro + * @description Use a Qwen3 encoder as a causal LLM to enhance a prompt, then encode it for Z-Image and Flux Klein. + * + * Accepts the same Qwen3 encoder model used by Z-Image and Flux2 Klein text encoder nodes. + * Outputs the enhanced prompt as text plus ready-to-use conditioning for both pipelines. */ - MetadataToIntegerInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + Qwen3PromptProInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20226,99 +34159,112 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} - */ - label?: "* CUSTOM LABEL *" | "width" | "height" | "seed" | "steps" | "clip_skip" | "cfg_scale_start_step" | "cfg_scale_end_step"; - /** - * Custom Label - * @description Label for this metadata item - * @default null + * Prompt + * @description Input text prompt. + * @default */ - custom_label?: string | null; + prompt?: string; /** - * Default Value - * @description The default integer to use if not found in the metadata - * @default null + * System Prompt + * @description System prompt that guides prompt enhancement (generation step). + * @default You are an expert prompt writer for AI image generation. Given a brief description, expand it into a detailed, vivid prompt suitable for generating high-quality images. Only output the expanded prompt, nothing else. */ - default_value?: number | null; + system_prompt?: string; /** - * type - * @default metadata_to_integer - * @constant + * Conditioning System Prompt + * @description System prompt prepended when encoding the enhanced prompt into conditioning. Leave empty for default behavior. + * @default */ - type: "metadata_to_integer"; - }; - /** - * Metadata To LoRA Collection - * @description Extracts Lora(s) from metadata into a collection - */ - MetadataToLorasCollectionInvocation: { + conditioning_system_prompt?: string; /** - * @description Optional metadata to be saved with the image + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description A mask defining the region that the conditioning applies to. + * @default null */ - id: string; + mask?: components["schemas"]["TensorField"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Max Tokens + * @description Maximum number of tokens to generate. + * @default 300 */ - is_intermediate?: boolean; + max_tokens?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Temperature + * @description Sampling temperature. Higher = more creative, lower = more focused. + * @default 0.7 */ - use_cache?: boolean; + temperature?: number; /** - * Custom Label - * @description Label for this metadata item - * @default loras + * Top P + * @description Nucleus sampling threshold. + * @default 0.9 */ - custom_label?: string; + top_p?: number; /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. - * @default [] + * Enable Thinking + * @description Enable Qwen3 thinking mode during generation. Slower but potentially higher quality. + * @default false */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; + enable_thinking?: boolean; /** * type - * @default metadata_to_lora_collection + * @default qwen3_prompt_pro * @constant */ - type: "metadata_to_lora_collection"; + type: "qwen3_prompt_pro"; }; /** - * MetadataToLorasCollectionOutput - * @description Model loader output + * Qwen3PromptProOutput + * @description Output containing the enhanced prompt text and conditioning for Z-Image and Flux Klein. */ - MetadataToLorasCollectionOutput: { + Qwen3PromptProOutput: { /** - * LoRAs - * @description Collection of LoRA model and weights + * Enhanced Prompt + * @description The enhanced prompt text */ - lora: components["schemas"]["LoRAField"][]; + enhanced_prompt: string; + /** @description Z-Image conditioning embeddings */ + z_image_conditioning: components["schemas"]["ZImageConditioningField"]; + /** + * Z Image Prompt + * @description The prompt text encoded for Z-Image conditioning + */ + z_image_prompt: string; + /** @description Flux Klein conditioning embeddings */ + flux_conditioning: components["schemas"]["FluxConditioningField"]; + /** + * Flux Prompt + * @description The prompt text encoded for Flux Klein conditioning + */ + flux_prompt: string; /** * type - * @default metadata_to_lora_collection_output + * @default qwen3_prompt_pro_output * @constant */ - type: "metadata_to_lora_collection_output"; + type: "qwen3_prompt_pro_output"; }; /** - * Metadata To LoRAs - * @description Extracts a Loras value of a label from metadata + * Qwen3VariantType + * @description Qwen3 text encoder variants based on model size. + * @enum {string} */ - MetadataToLorasInvocation: { + Qwen3VariantType: "qwen3_4b" | "qwen3_8b"; + /** + * RGB Merge + * @description Merge RGB color channels and alpha + */ + RGBMergeInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -20342,29 +34288,42 @@ export type components = { */ use_cache?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) + * @description The red channel * @default null */ - unet?: components["schemas"]["UNetField"] | null; + r_channel?: components["schemas"]["ImageField"] | null; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @description The green channel * @default null */ - clip?: components["schemas"]["CLIPField"] | null; + g_channel?: components["schemas"]["ImageField"] | null; + /** + * @description The blue channel + * @default null + */ + b_channel?: components["schemas"]["ImageField"] | null; + /** + * @description The alpha channel + * @default null + */ + alpha_channel?: components["schemas"]["ImageField"] | null; /** * type - * @default metadata_to_loras + * @default rgb_merge * @constant */ - type: "metadata_to_loras"; + type: "rgb_merge"; }; /** - * Metadata To Model - * @description Extracts a Model value of a label from metadata + * RGB Split + * @description Split an image into RGB color channels and alpha */ - MetadataToModelInvocation: { + RGBSplitInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -20388,77 +34347,52 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default model - * @enum {string} - */ - label?: "* CUSTOM LABEL *" | "model"; - /** - * Custom Label - * @description Label for this metadata item - * @default null - */ - custom_label?: string | null; - /** - * @description The default model to use if not found in the metadata + * @description The image to split into channels * @default null */ - default_value?: components["schemas"]["ModelIdentifierField"] | null; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default metadata_to_model + * @default rgb_split * @constant */ - type: "metadata_to_model"; + type: "rgb_split"; }; /** - * MetadataToModelOutput - * @description String to main model output + * RGBSplitOutput + * @description Base class for invocations that output three L-mode images (R, G, B) */ - MetadataToModelOutput: { - /** - * Model - * @description Main model (UNet, VAE, CLIP) to load - */ - model: components["schemas"]["ModelIdentifierField"]; - /** - * Name - * @description Model Name - */ - name: string; - /** - * UNet - * @description UNet (scheduler, LoRAs) - */ - unet: components["schemas"]["UNetField"]; + RGBSplitOutput: { + /** @description Grayscale image of the red channel */ + r_channel: components["schemas"]["ImageField"]; + /** @description Grayscale image of the green channel */ + g_channel: components["schemas"]["ImageField"]; + /** @description Grayscale image of the blue channel */ + b_channel: components["schemas"]["ImageField"]; + /** @description Grayscale image of the alpha channel */ + alpha_channel: components["schemas"]["ImageField"]; /** - * VAE - * @description VAE + * Width + * @description The width of the image in pixels */ - vae: components["schemas"]["VAEField"]; + width: number; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Height + * @description The height of the image in pixels */ - clip: components["schemas"]["CLIPField"]; + height: number; /** * type - * @default metadata_to_model_output + * @default rgb_split_output * @constant */ - type: "metadata_to_model_output"; + type: "rgb_split_output"; }; /** - * Metadata To SDXL LoRAs - * @description Extracts a SDXL Loras value of a label from metadata + * Random Float + * @description Outputs a single random float */ - MetadataToSDXLLorasInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + RandomFloatInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20473,44 +34407,39 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Low + * @description The inclusive low value + * @default 0 */ - unet?: components["schemas"]["UNetField"] | null; + low?: number; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * High + * @description The exclusive high value + * @default 1 */ - clip?: components["schemas"]["CLIPField"] | null; + high?: number; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Decimals + * @description The number of decimal places to round to + * @default 2 */ - clip2?: components["schemas"]["CLIPField"] | null; + decimals?: number; /** * type - * @default metadata_to_sdlx_loras + * @default rand_float * @constant */ - type: "metadata_to_sdlx_loras"; + type: "rand_float"; }; /** - * Metadata To SDXL Model - * @description Extracts a SDXL Model value of a label from metadata + * Random Image Size + * @description Outputs a random size from a list of sizes. */ - MetadataToSDXLModelInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + RandomImageSizeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20529,82 +34458,49 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default model - * @enum {string} + * Seed + * @description A seed for the randomness. Use -1 for non-deterministic. + * @default -1 */ - label?: "* CUSTOM LABEL *" | "model"; + seed?: number; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Size List + * @default 1024x1024,888x1184,1184x888,840x1256,1256x840,768x1368,1368x768 */ - custom_label?: string | null; + size_list?: string; /** - * @description The default SDXL Model to use if not found in the metadata - * @default null + * Multiples of 32 + * @default false */ - default_value?: components["schemas"]["ModelIdentifierField"] | null; + multiples_of_32?: boolean; /** * type - * @default metadata_to_sdxl_model + * @default random_image_size_invocation * @constant */ - type: "metadata_to_sdxl_model"; + type: "random_image_size_invocation"; }; /** - * MetadataToSDXLModelOutput - * @description String to SDXL main model output + * RandomImageSizeOutput + * @description Random Image Size Output */ - MetadataToSDXLModelOutput: { - /** - * Model - * @description Main model (UNet, VAE, CLIP) to load - */ - model: components["schemas"]["ModelIdentifierField"]; - /** - * Name - * @description Model Name - */ - name: string; - /** - * UNet - * @description UNet (scheduler, LoRAs) - */ - unet: components["schemas"]["UNetField"]; - /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - */ - clip: components["schemas"]["CLIPField"]; - /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - */ - clip2: components["schemas"]["CLIPField"]; - /** - * VAE - * @description VAE - */ - vae: components["schemas"]["VAEField"]; + RandomImageSizeOutput: { + /** Width */ + width: number; + /** Height */ + height: number; /** * type - * @default metadata_to_sdxl_model_output + * @default random_image_size * @constant */ - type: "metadata_to_sdxl_model_output"; + type: "random_image_size"; }; /** - * Metadata To Scheduler - * @description Extracts a Scheduler value of a label from metadata + * Random Integer + * @description Outputs a single random integer. */ - MetadataToSchedulerInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + RandomIntInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20619,46 +34515,33 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default scheduler - * @enum {string} - */ - label?: "* CUSTOM LABEL *" | "scheduler"; - /** - * Custom Label - * @description Label for this metadata item - * @default null + * Low + * @description The inclusive low value + * @default 0 */ - custom_label?: string | null; + low?: number; /** - * Default Value - * @description The default scheduler to use if not found in the metadata - * @default euler - * @enum {string} + * High + * @description The exclusive high value + * @default 2147483647 */ - default_value?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + high?: number; /** * type - * @default metadata_to_scheduler + * @default rand_int * @constant */ - type: "metadata_to_scheduler"; + type: "rand_int"; }; /** - * Metadata To String Collection - * @description Extracts a string collection value of a label from metadata + * Random LoRA Mixer + * @description Returns a random Collection of LoRAs selected from an input Collection. */ - MetadataToStringCollectionInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + RandomLoRAMixerInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20673,45 +34556,96 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * LoRAs + * @description A random selection of the input LoRAs with random weights + * @default null */ - label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt"; + loras?: components["schemas"]["LoRAField"][] | null; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Seed + * @description A seed to use for the random number generator. This allows reproducability. Leave as -1 for non-deterministic + * @default -1 */ - custom_label?: string | null; + seed?: number; /** - * Default Value - * @description The default string collection to use if not found in the metadata - * @default null + * Min LoRAs + * @description The mininum number of LoRAs to output + * @default 1 */ - default_value?: string[] | null; + min_loras?: number; + /** + * Max LoRAs + * @description The maximum number of LoRAs to output + * @default 3 + */ + max_loras?: number; + /** + * Min Weight + * @description The mininum weight to apply to a LoRA + * @default 0.05 + */ + min_weight?: number; + /** + * Max Weight + * @description The maxinum weight to apply to a LoRA + * @default 1 + */ + max_weight?: number; + /** + * Trigger Word Delimiter + * @description A character sequence to place between each trigger word + * @default , + */ + trigger_word_delimiter?: string; + /** + * LoRA Names Delimiter + * @description A character sequence to place between each LoRA name + * @default , + */ + lora_names_delimiter?: string; /** * type - * @default metadata_to_string_collection + * @default random_lora_mixer_invocation * @constant */ - type: "metadata_to_string_collection"; + type: "random_lora_mixer_invocation"; }; /** - * Metadata To String - * @description Extracts a string value of a label from metadata + * RandomLoRAMixerOutput + * @description Random LoRA Mixer Output */ - MetadataToStringInvocation: { + RandomLoRAMixerOutput: { /** - * @description Optional metadata to be saved with the image - * @default null + * LoRAs + * @description A random selection of the input LoRAs with random weights */ - metadata?: components["schemas"]["MetadataField"] | null; + loras: components["schemas"]["LoRAField"][]; + /** + * Trigger Words + * @description A string with delimited trigger words for the random LoRAs + */ + trigger_words: string; + /** + * Lora Names + * @description A string with the names and weights of all applied LoRAs + */ + lora_names: string; + /** + * type + * @default random_lora_mixer_output + * @constant + */ + type: "random_lora_mixer_output"; + }; + /** + * Random Range + * @description Creates a collection of random numbers + */ + RandomRangeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20726,45 +34660,45 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Low + * @description The inclusive low value + * @default 0 */ - label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt"; + low?: number; /** - * Custom Label - * @description Label for this metadata item - * @default null + * High + * @description The exclusive high value + * @default 2147483647 */ - custom_label?: string | null; + high?: number; /** - * Default Value - * @description The default string to use if not found in the metadata - * @default null + * Size + * @description The number of values to generate + * @default 1 */ - default_value?: string | null; + size?: number; + /** + * Seed + * @description The seed for the RNG (omit for random) + * @default 0 + */ + seed?: number; /** * type - * @default metadata_to_string + * @default random_range * @constant */ - type: "metadata_to_string"; + type: "random_range"; }; /** - * Metadata To T2I-Adapters - * @description Extracts a T2I-Adapters value of a label from metadata + * Integer Range + * @description Creates a range of numbers from start to stop with step */ - MetadataToT2IAdaptersInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + RangeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20783,28 +34717,35 @@ export type components = { */ use_cache?: boolean; /** - * T2I-Adapter - * @description IP-Adapter to apply - * @default null + * Start + * @description The start of the range + * @default 0 */ - t2i_adapter_list?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + start?: number; /** - * type - * @default metadata_to_t2i_adapters - * @constant + * Stop + * @description The stop of the range + * @default 10 */ - type: "metadata_to_t2i_adapters"; - }; - /** - * Metadata To VAE - * @description Extracts a VAE value of a label from metadata - */ - MetadataToVAEInvocation: { + stop?: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Step + * @description The step of the range + * @default 1 + */ + step?: number; + /** + * type + * @default range + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "range"; + }; + /** + * Integer Range of Size + * @description Creates a range from start to start + (size * step) incremented by step + */ + RangeOfSizeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20823,70 +34764,35 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default vae - * @enum {string} + * Start + * @description The start of the range + * @default 0 */ - label?: "* CUSTOM LABEL *" | "vae"; + start?: number; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Size + * @description The number of values + * @default 1 */ - custom_label?: string | null; + size?: number; /** - * @description The default VAE to use if not found in the metadata - * @default null + * Step + * @description The step of the range + * @default 1 */ - default_value?: components["schemas"]["VAEField"] | null; + step?: number; /** * type - * @default metadata_to_vae + * @default range_of_size * @constant */ - type: "metadata_to_vae"; - }; - /** - * ModelFormat - * @description Storage format of model. - * @enum {string} - */ - ModelFormat: "omi" | "diffusers" | "checkpoint" | "lycoris" | "onnx" | "olive" | "embedding_file" | "embedding_folder" | "invokeai" | "t5_encoder" | "qwen3_encoder" | "bnb_quantized_int8b" | "bnb_quantized_nf4b" | "gguf_quantized" | "unknown"; - /** ModelIdentifierField */ - ModelIdentifierField: { - /** - * Key - * @description The model's unique key - */ - key: string; - /** - * Hash - * @description The model's BLAKE3 hash - */ - hash: string; - /** - * Name - * @description The model's name - */ - name: string; - /** @description The model's base model type */ - base: components["schemas"]["BaseModelType"]; - /** @description The model's type */ - type: components["schemas"]["ModelType"]; - /** - * @description The submodel to load, if this is a main model - * @default null - */ - submodel_type?: components["schemas"]["SubModelType"] | null; + type: "range_of_size"; }; /** - * Any Model - * @description Selects any model, outputting it its identifier. Be careful with this one! The identifier will be accepted as - * input for any model, even if the model types don't match. If you connect this to a mismatched input, you'll get an - * error. + * Reapply LoRA Weight + * @description Allows specifying an alternate weight for a LoRA after it has been selected with a LoRA selector node */ - ModelIdentifierInvocation: { + ReapplyLoRAWeightInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -20905,544 +34811,562 @@ export type components = { */ use_cache?: boolean; /** - * Model - * @description The model to select + * LoRA * @default null */ - model?: components["schemas"]["ModelIdentifierField"] | null; + lora?: components["schemas"]["LoRAField"] | null; + /** + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 + */ + weight?: number; /** * type - * @default model_identifier + * @default reapply_lora_weight_invocation * @constant */ - type: "model_identifier"; + type: "reapply_lora_weight_invocation"; }; /** - * ModelIdentifierOutput - * @description Model identifier output + * ReapplyLoRAWeightOutput + * @description Reapply LoRA Weight Output */ - ModelIdentifierOutput: { + ReapplyLoRAWeightOutput: { /** - * Model - * @description Model identifier + * LoRA + * @description LoRA model from a LoRA Selector node */ - model: components["schemas"]["ModelIdentifierField"]; + lora: components["schemas"]["LoRAField"]; /** * type - * @default model_identifier_output + * @default reapply_lora_weight_output * @constant */ - type: "model_identifier_output"; + type: "reapply_lora_weight_output"; }; /** - * ModelInstallCancelledEvent - * @description Event model for model_install_cancelled + * RecallParameter + * @description Request model for updating recallable parameters. */ - ModelInstallCancelledEvent: { + RecallParameter: { /** - * Timestamp - * @description The timestamp of the event + * Positive Prompt + * @description Positive prompt text */ - timestamp: number; + positive_prompt?: string | null; /** - * Id - * @description The ID of the install job + * Negative Prompt + * @description Negative prompt text */ - id: number; + negative_prompt?: string | null; /** - * Source - * @description Source of the model; local path, repo_id or url + * Model + * @description Main model name/identifier */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; - }; - /** - * ModelInstallCompleteEvent - * @description Event model for model_install_complete - */ - ModelInstallCompleteEvent: { + model?: string | null; /** - * Timestamp - * @description The timestamp of the event + * Refiner Model + * @description Refiner model name/identifier */ - timestamp: number; + refiner_model?: string | null; /** - * Id - * @description The ID of the install job + * Vae Model + * @description VAE model name/identifier */ - id: number; + vae_model?: string | null; /** - * Source - * @description Source of the model; local path, repo_id or url + * Scheduler + * @description Scheduler name */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + scheduler?: string | null; /** - * Key - * @description Model config record key + * Steps + * @description Number of generation steps */ - key: string; + steps?: number | null; /** - * Total Bytes - * @description Size of the model (may be None for installation of a local path) + * Refiner Steps + * @description Number of refiner steps */ - total_bytes: number | null; + refiner_steps?: number | null; /** - * Config - * @description The installed model's config + * Cfg Scale + * @description CFG scale for guidance */ - config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; - }; - /** - * ModelInstallDownloadProgressEvent - * @description Event model for model_install_download_progress - */ - ModelInstallDownloadProgressEvent: { + cfg_scale?: number | null; /** - * Timestamp - * @description The timestamp of the event + * Cfg Rescale Multiplier + * @description CFG rescale multiplier */ - timestamp: number; + cfg_rescale_multiplier?: number | null; /** - * Id - * @description The ID of the install job + * Refiner Cfg Scale + * @description Refiner CFG scale */ - id: number; + refiner_cfg_scale?: number | null; /** - * Source - * @description Source of the model; local path, repo_id or url + * Guidance + * @description Guidance scale */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + guidance?: number | null; /** - * Local Path - * @description Where model is downloading to + * Width + * @description Image width in pixels */ - local_path: string; + width?: number | null; /** - * Bytes - * @description Number of bytes downloaded so far + * Height + * @description Image height in pixels */ - bytes: number; + height?: number | null; /** - * Total Bytes - * @description Total size of download, including all files + * Seed + * @description Random seed */ - total_bytes: number; + seed?: number | null; /** - * Parts - * @description Progress of downloading URLs that comprise the model, if any + * Denoise Strength + * @description Denoising strength */ - parts: { - [key: string]: number | string; - }[]; - }; - /** - * ModelInstallDownloadStartedEvent - * @description Event model for model_install_download_started - */ - ModelInstallDownloadStartedEvent: { + denoise_strength?: number | null; /** - * Timestamp - * @description The timestamp of the event + * Refiner Denoise Start + * @description Refiner denoising start */ - timestamp: number; + refiner_denoise_start?: number | null; /** - * Id - * @description The ID of the install job + * Clip Skip + * @description CLIP skip layers */ - id: number; + clip_skip?: number | null; /** - * Source - * @description Source of the model; local path, repo_id or url + * Seamless X + * @description Enable seamless X tiling */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + seamless_x?: boolean | null; /** - * Local Path - * @description Where model is downloading to + * Seamless Y + * @description Enable seamless Y tiling */ - local_path: string; + seamless_y?: boolean | null; /** - * Bytes - * @description Number of bytes downloaded so far + * Refiner Positive Aesthetic Score + * @description Refiner positive aesthetic score + */ + refiner_positive_aesthetic_score?: number | null; + /** + * Refiner Negative Aesthetic Score + * @description Refiner negative aesthetic score + */ + refiner_negative_aesthetic_score?: number | null; + /** + * Loras + * @description List of LoRAs with their weights */ - bytes: number; + loras?: components["schemas"]["LoRARecallParameter"][] | null; /** - * Total Bytes - * @description Total size of download, including all files + * Control Layers + * @description List of control adapters (ControlNet, T2I Adapter, Control LoRA) with their settings */ - total_bytes: number; + control_layers?: components["schemas"]["ControlNetRecallParameter"][] | null; /** - * Parts - * @description Progress of downloading URLs that comprise the model, if any + * Ip Adapters + * @description List of IP Adapters with their settings */ - parts: { - [key: string]: number | string; - }[]; + ip_adapters?: components["schemas"]["IPAdapterRecallParameter"][] | null; }; /** - * ModelInstallDownloadsCompleteEvent - * @description Emitted once when an install job becomes active. + * RecallParametersUpdatedEvent + * @description Event model for recall_parameters_updated */ - ModelInstallDownloadsCompleteEvent: { + RecallParametersUpdatedEvent: { /** * Timestamp * @description The timestamp of the event */ timestamp: number; /** - * Id - * @description The ID of the install job + * Queue Id + * @description The ID of the queue */ - id: number; + queue_id: string; /** - * Source - * @description Source of the model; local path, repo_id or url + * Parameters + * @description The recall parameters that were updated */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + parameters: { + [key: string]: unknown; + }; }; /** - * ModelInstallErrorEvent - * @description Event model for model_install_error + * Create Rectangle Mask + * @description Create a rectangular mask. */ - ModelInstallErrorEvent: { + RectangleMaskInvocation: { /** - * Timestamp - * @description The timestamp of the event + * @description Optional metadata to be saved with the image + * @default null */ - timestamp: number; + metadata?: components["schemas"]["MetadataField"] | null; /** * Id - * @description The ID of the install job + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - id: number; + id: string; /** - * Source - * @description Source of the model; local path, repo_id or url + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + is_intermediate?: boolean; /** - * Error Type - * @description The name of the exception + * Use Cache + * @description Whether or not to use the cache + * @default true */ - error_type: string; + use_cache?: boolean; /** - * Error - * @description A text description of the exception + * Width + * @description The width of the entire mask. + * @default null */ - error: string; - }; - /** - * ModelInstallJob - * @description Object that tracks the current status of an install request. - */ - ModelInstallJob: { + width?: number | null; /** - * Id - * @description Unique ID for this job + * Height + * @description The height of the entire mask. + * @default null */ - id: number; + height?: number | null; /** - * @description Current status of install process - * @default waiting + * X Left + * @description The left x-coordinate of the rectangular masked region (inclusive). + * @default null */ - status?: components["schemas"]["InstallStatus"]; + x_left?: number | null; /** - * Error Reason - * @description Information about why the job failed + * Y Top + * @description The top y-coordinate of the rectangular masked region (inclusive). + * @default null */ - error_reason?: string | null; - /** @description Configuration information (e.g. 'description') to apply to model. */ - config_in?: components["schemas"]["ModelRecordChanges"]; + y_top?: number | null; /** - * Config Out - * @description After successful installation, this will hold the configuration object. + * Rectangle Width + * @description The width of the rectangular masked region. + * @default null */ - config_out?: (components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]) | null; + rectangle_width?: number | null; /** - * Inplace - * @description Leave model in its current location; otherwise install under models directory - * @default false + * Rectangle Height + * @description The height of the rectangular masked region. + * @default null */ - inplace?: boolean; + rectangle_height?: number | null; /** - * Source - * @description Source (URL, repo_id, or local path) of model + * type + * @default rectangle_mask + * @constant */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + type: "rectangle_mask"; + }; + /** + * RemoteModelFile + * @description Information about a downloadable file that forms part of a model. + */ + RemoteModelFile: { /** - * Local Path - * Format: path - * @description Path to locally-downloaded model; may be the same as the source + * Url + * Format: uri + * @description The url to download this model file */ - local_path: string; + url: string; /** - * Bytes - * @description For a remote model, the number of bytes downloaded so far (may not be available) - * @default 0 + * Path + * Format: path + * @description The path to the file, relative to the model root */ - bytes?: number; + path: string; /** - * Total Bytes - * @description Total size of the model to be installed + * Size + * @description The size of this file, in bytes * @default 0 */ - total_bytes?: number; - /** - * Source Metadata - * @description Metadata provided by the model source - */ - source_metadata?: (components["schemas"]["BaseMetadata"] | components["schemas"]["HuggingFaceMetadata"]) | null; + size?: number | null; /** - * Download Parts - * @description Download jobs contributing to this install + * Sha256 + * @description SHA256 hash of this model (not always available) */ - download_parts?: components["schemas"]["DownloadJob"][]; + sha256?: string | null; + }; + /** RemoveImagesFromBoardResult */ + RemoveImagesFromBoardResult: { /** - * Error - * @description On an error condition, this field will contain the text of the exception + * Affected Boards + * @description The ids of boards affected by the delete operation */ - error?: string | null; + affected_boards: string[]; /** - * Error Traceback - * @description On an error condition, this field will contain the exception traceback + * Removed Images + * @description The image names that were removed from their board */ - error_traceback?: string | null; + removed_images: string[]; }; /** - * ModelInstallStartedEvent - * @description Event model for model_install_started + * Resize Latents + * @description Resizes latents to explicit width/height (in pixels). Provided dimensions are floor-divided by 8. */ - ModelInstallStartedEvent: { + ResizeLatentsInvocation: { /** - * Timestamp - * @description The timestamp of the event + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - timestamp: number; + id: string; /** - * Id - * @description The ID of the install job + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - id: number; + is_intermediate?: boolean; /** - * Source - * @description Source of the model; local path, repo_id or url + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + use_cache?: boolean; + /** + * @description Latents tensor + * @default null + */ + latents?: components["schemas"]["LatentsField"] | null; + /** + * Width + * @description Width of output (px) + * @default null + */ + width?: number | null; + /** + * Height + * @description Width of output (px) + * @default null + */ + height?: number | null; + /** + * Mode + * @description Interpolation mode + * @default bilinear + * @enum {string} + */ + mode?: "nearest" | "linear" | "bilinear" | "bicubic" | "trilinear" | "area" | "nearest-exact"; + /** + * Antialias + * @description Whether or not to apply antialiasing (bilinear or bicubic only) + * @default false + */ + antialias?: boolean; + /** + * type + * @default lresize + * @constant + */ + type: "lresize"; }; /** - * ModelLoadCompleteEvent - * @description Event model for model_load_complete + * ResourceOrigin + * @description The origin of a resource (eg image). + * + * - INTERNAL: The resource was created by the application. + * - EXTERNAL: The resource was not created by the application. + * This may be a user-initiated upload, or an internal application upload (eg Canvas init image). + * @enum {string} */ - ModelLoadCompleteEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; + ResourceOrigin: "internal" | "external"; + /** Retrieve Images from Board */ + RetrieveBoardImagesInvocation: { /** - * Config - * @description The model's config + * @description Optional metadata to be saved with the image + * @default null */ - config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; + metadata?: components["schemas"]["MetadataField"] | null; /** - * @description The submodel type, if any - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - submodel_type: components["schemas"]["SubModelType"] | null; - }; - /** - * ModelLoadStartedEvent - * @description Event model for model_load_started - */ - ModelLoadStartedEvent: { + id: string; /** - * Timestamp - * @description The timestamp of the event + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - timestamp: number; + is_intermediate?: boolean; /** - * Config - * @description The model's config + * Use Cache + * @description Whether or not to use the cache + * @default false */ - config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; + use_cache?: boolean; /** - * @description The submodel type, if any + * @description Input board containing images to be retrieved. If not provided or set to None, it will grab images from uncategorized. * @default null */ - submodel_type: components["schemas"]["SubModelType"] | null; - }; - /** - * ModelLoaderOutput - * @description Model loader output - */ - ModelLoaderOutput: { + input_board?: components["schemas"]["BoardField"] | null; /** - * VAE - * @description VAE + * Num Images + * @description Number of images to retrieve: specify 'all', a single index like '10', specific indices like '1,4,6', or a range like '30-50'. + * @default all */ - vae: components["schemas"]["VAEField"]; + num_images?: string; /** - * type - * @default model_loader_output - * @constant + * Category + * @description Category of images to retrieve; select either 'images' or 'assets'. + * @default images + * @enum {string} */ - type: "model_loader_output"; + category?: "images" | "assets"; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Starred Only + * @description Retrieve only starred images if set to True. + * @default false */ - clip: components["schemas"]["CLIPField"]; + starred_only?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) + * Keyword + * @description Keyword to filter images by metadata. Only images with metadata containing this keyword will be retrieved. + * @default null */ - unet: components["schemas"]["UNetField"]; + keyword?: string | null; + /** + * type + * @default Retrieve_Board_Images + * @constant + */ + type: "Retrieve_Board_Images"; }; /** - * ModelRecordChanges - * @description A set of changes to apply to a model. + * Retrieve Flux Conditioning + * @description Retrieves one or more FLUX Conditioning objects (CLIP and T5 embeddings) + * from an SQLite database using unique identifiers. + * Outputs a single selected conditioning and a list of all retrieved conditionings. + * Includes an option to update the timestamp of retrieved entries. + * The input conditioning IDs are sorted for deterministic retrieval. */ - ModelRecordChanges: { - /** - * Source - * @description original source of the model - */ - source?: string | null; - /** @description type of model source */ - source_type?: components["schemas"]["ModelSourceType"] | null; + RetrieveFluxConditioningInvocation: { /** - * Source Api Response - * @description metadata from remote source - */ - source_api_response?: string | null; - /** - * Name - * @description Name of the model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - name?: string | null; + id: string; /** - * Path - * @description Path to the model. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - path?: string | null; + is_intermediate?: boolean; /** - * Description - * @description Model description + * Use Cache + * @description Whether or not to use the cache + * @default true */ - description?: string | null; - /** @description The base model. */ - base?: components["schemas"]["BaseModelType"] | null; - /** @description Type of model */ - type?: components["schemas"]["ModelType"] | null; + use_cache?: boolean; /** - * Key - * @description Database ID for this model + * Conditioning Id Or List + * @description The unique identifier(s) of the Flux Conditioning(s) to retrieve. + * @default null */ - key?: string | null; + conditioning_id_or_list?: string | string[] | null; /** - * Hash - * @description hash of model file + * Select Index + * @description Index of the retrieved conditioning to output as the single 'conditioning' field. If out of bounds, uses modulus. + * @default 0 */ - hash?: string | null; + select_index?: number; /** - * File Size - * @description Size of model file + * Touch Timestamp + * @description When true, updates the timestamp of retrieved entries to 'now', preventing early purge. + * @default false */ - file_size?: number | null; + touch_timestamp?: boolean; /** - * Format - * @description format of model file + * type + * @default retrieve_flux_conditioning + * @constant */ - format?: string | null; + type: "retrieve_flux_conditioning"; + }; + /** + * RetrieveFluxConditioningMultiOutput + * @description Output for the Retrieve Flux Conditioning node, providing a single selected + * conditioning and a list of all retrieved conditionings. + */ + RetrieveFluxConditioningMultiOutput: { + /** @description A single selected Flux Conditioning (selected by index from the retrieved list) */ + conditioning: components["schemas"]["FluxConditioningField"]; /** - * Trigger Phrases - * @description Set of trigger phrases for this model + * Conditioning List + * @description A list of all retrieved Flux Conditionings */ - trigger_phrases?: string[] | null; + conditioning_list: components["schemas"]["FluxConditioningField"][]; /** - * Default Settings - * @description Default settings for this model + * type + * @default retrieve_flux_conditioning_multi_output + * @constant */ - default_settings?: components["schemas"]["MainModelDefaultSettings"] | components["schemas"]["LoraModelDefaultSettings"] | components["schemas"]["ControlAdapterDefaultSettings"] | null; + type: "retrieve_flux_conditioning_multi_output"; + }; + /** + * Bitize + * @description Crush an image to one-bit pixels + */ + RetroBitizeInvocation: { /** - * Cpu Only - * @description Whether this model should run on CPU only + * @description Optional metadata to be saved with the image + * @default null */ - cpu_only?: boolean | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Variant - * @description The variant of the model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - variant?: components["schemas"]["ModelVariantType"] | components["schemas"]["ClipVariantType"] | components["schemas"]["FluxVariantType"] | components["schemas"]["Flux2VariantType"] | components["schemas"]["ZImageVariantType"] | components["schemas"]["Qwen3VariantType"] | null; - /** @description The prediction type of the model. */ - prediction_type?: components["schemas"]["SchedulerPredictionType"] | null; + id: string; /** - * Upcast Attention - * @description Whether to upcast attention. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - upcast_attention?: boolean | null; + is_intermediate?: boolean; /** - * Config Path - * @description Path to config file for model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - config_path?: string | null; - }; - /** ModelRelationshipBatchRequest */ - ModelRelationshipBatchRequest: { + use_cache?: boolean; /** - * Model Keys - * @description List of model keys to fetch related models for + * @description Input image for pixelization + * @default null */ - model_keys: string[]; - }; - /** ModelRelationshipCreateRequest */ - ModelRelationshipCreateRequest: { + image?: components["schemas"]["ImageField"] | null; /** - * Model Key 1 - * @description The key of the first model in the relationship + * Dither + * @description Dither the Bitized image + * @default true */ - model_key_1: string; + dither?: boolean; /** - * Model Key 2 - * @description The key of the second model in the relationship + * type + * @default retro_bitize + * @constant */ - model_key_2: string; - }; - /** - * ModelRepoVariant - * @description Various hugging face variants on the diffusers format. - * @enum {string} - */ - ModelRepoVariant: "" | "fp16" | "fp32" | "onnx" | "openvino" | "flax"; - /** - * ModelSourceType - * @description Model source type. - * @enum {string} - */ - ModelSourceType: "path" | "url" | "hf_repo_id"; - /** - * ModelType - * @description Model type. - * @enum {string} - */ - ModelType: "onnx" | "main" | "vae" | "lora" | "control_lora" | "controlnet" | "embedding" | "ip_adapter" | "clip_vision" | "clip_embed" | "t2i_adapter" | "t5_encoder" | "qwen3_encoder" | "spandrel_image_to_image" | "siglip" | "flux_redux" | "llava_onevision" | "unknown"; - /** - * ModelVariantType - * @description Variant type. - * @enum {string} - */ - ModelVariantType: "normal" | "inpaint" | "depth"; - /** - * ModelsList - * @description Return list of configs. - */ - ModelsList: { - /** Models */ - models: (components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"])[]; + type: "retro_bitize"; }; /** - * Multiply Integers - * @description Multiplies two numbers + * CRT + * @description Distort the input image, simulating CRT display curvature */ - MultiplyInvocation: { + RetroCRTCurvatureInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -21461,47 +35385,69 @@ export type components = { */ use_cache?: boolean; /** - * A - * @description The first number - * @default 0 + * @description Input image for pixelization + * @default null */ - a?: number; + image?: components["schemas"]["ImageField"] | null; /** - * B - * @description The second number - * @default 0 + * Crt Width + * @description Horizontal resolution of the CRT; smaller = more noticeable + * @default 240 */ - b?: number; + crt_width?: number; /** - * type - * @default mul - * @constant + * Crt Height + * @description Vertical resolution of the CRT; smaller = more noticeable + * @default 160 */ - type: "mul"; - }; - /** NodeFieldValue */ - NodeFieldValue: { + crt_height?: number; /** - * Node Path - * @description The node into which this batch data item will be substituted. + * Crt Curvature + * @description Curvature factor; smaller = stronger inward curve + * @default 3 */ - node_path: string; + crt_curvature?: number; /** - * Field Name - * @description The field into which this batch data item will be substituted. + * Scanlines Opacity + * @description Opacity of CRT scan lines + * @default 1 */ - field_name: string; + scanlines_opacity?: number; /** - * Value - * @description The value to substitute into the node/field. + * Vignette Opacity + * @description Vignette opacity + * @default 0.5 + */ + vignette_opacity?: number; + /** + * Vignette Roundness + * @description Vignette opacity + * @default 5 + */ + vignette_roundness?: number; + /** + * Crt Brightness + * @description Factor by which to brighten the image if using scan lines + * @default 1.2 + */ + crt_brightness?: number; + /** + * type + * @default retro_crt_curvature + * @constant */ - value: string | number | components["schemas"]["ImageField"]; + type: "retro_crt_curvature"; }; /** - * Create Latent Noise - * @description Generates latent noise. + * Get Palette (Advanced) + * @description Get palette from an image, 256 colors max. Optionally export to a user-defined location. */ - NoiseInvocation: { + RetroGetPaletteAdvInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -21520,70 +35466,79 @@ export type components = { */ use_cache?: boolean; /** - * Seed - * @description Seed for random number generation - * @default 0 + * @description Input image to grab a palette from + * @default null */ - seed?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Width - * @description Width of output (px) - * @default 512 + * Export + * @description Save palette PNG to specified path with optional name + * @default true */ - width?: number; + export?: boolean; /** - * Height - * @description Height of output (px) - * @default 512 + * Subfolder + * @description Subfolder for the palette in nodes/Retroize/palettes/ folder + * @default */ - height?: number; + subfolder?: string; /** - * Use Cpu - * @description Use CPU for noise generation (for reproducible results across platforms) - * @default true + * Name + * @description Name for the palette image + * @default */ - use_cpu?: boolean; + name?: string; /** * type - * @default noise + * @default get_palette_adv * @constant */ - type: "noise"; + type: "get_palette_adv"; }; /** - * NoiseOutput - * @description Invocation noise output + * Get Palette + * @description Get palette from an image, 256 colors max. */ - NoiseOutput: { - /** @description Noise tensor */ - noise: components["schemas"]["LatentsField"]; + RetroGetPaletteInvocation: { /** - * Width - * @description Width of output (px) + * @description Optional metadata to be saved with the image + * @default null */ - width: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Height - * @description Height of output (px) + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - height: number; + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Input image to grab a palette from + * @default null + */ + image?: components["schemas"]["ImageField"] | null; /** * type - * @default noise_output + * @default get_palette * @constant */ - type: "noise_output"; + type: "get_palette"; }; /** - * Normal Map - * @description Generates a normal map. + * Palettize Advanced + * @description Palettize an image by applying a color palette */ - NormalMapInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + RetroPalettizeAdvInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -21607,119 +35562,116 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description Input image for pixelization * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * type - * @default normal_map - * @constant + * Palette Image + * @description Palette image + * @default null */ - type: "normal_map"; - }; - /** OffsetPaginatedResults[BoardDTO] */ - OffsetPaginatedResults_BoardDTO_: { + palette_image?: ("None" | "aap-64.png" | "atari-8-bit.png" | "commodore64.png" | "endesga-32.png" | "fantasy-24.png" | "mg\\mg-16-01.png" | "microsoft-windows.png" | "NES.png" | "nintendo-gameboy.png" | "pico-8.png" | "slso8.png") | null; /** - * Limit - * @description Limit of items to get + * Dither + * @description Apply dithering to image when palettizing + * @default false */ - limit: number; + dither?: boolean; /** - * Offset - * @description Offset from which to retrieve items + * Prequantize + * @description Apply 256-color quantization with specified method prior to applying the color palette + * @default false */ - offset: number; + prequantize?: boolean; /** - * Total - * @description Total number of items in result + * Quantizer + * @description Palettizer quantization method + * @default Fast Octree + * @enum {string} */ - total: number; + quantizer?: "Median Cut" | "Max Coverage" | "Fast Octree"; /** - * Items - * @description Items + * type + * @default retro_palettize_adv + * @constant */ - items: components["schemas"]["BoardDTO"][]; + type: "retro_palettize_adv"; }; - /** OffsetPaginatedResults[ImageDTO] */ - OffsetPaginatedResults_ImageDTO_: { + /** + * Palettize + * @description Palettize an image by applying a color palette + */ + RetroPalettizeInvocation: { /** - * Limit - * @description Limit of items to get + * @description Optional metadata to be saved with the image + * @default null */ - limit: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Offset - * @description Offset from which to retrieve items + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - offset: number; + id: string; /** - * Total - * @description Total number of items in result + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - total: number; + is_intermediate?: boolean; /** - * Items - * @description Items + * Use Cache + * @description Whether or not to use the cache + * @default true */ - items: components["schemas"]["ImageDTO"][]; - }; - /** - * OrphanedModelInfo - * @description Information about an orphaned model directory. - */ - OrphanedModelInfo: { + use_cache?: boolean; /** - * Path - * @description Relative path to the orphaned directory from models root + * @description Input image for pixelization + * @default null */ - path: string; + image?: components["schemas"]["ImageField"] | null; /** - * Absolute Path - * @description Absolute path to the orphaned directory + * @description Palette image + * @default null */ - absolute_path: string; + palette_image?: components["schemas"]["ImageField"] | null; /** - * Files - * @description List of model files in this directory + * Palette Path + * @description Palette image path, including ".png" extension + * @default */ - files: string[]; + palette_path?: string; /** - * Size Bytes - * @description Total size of all files in bytes + * Dither + * @description Apply dithering to image when palettizing + * @default false */ - size_bytes: number; - }; - /** - * OutputFieldJSONSchemaExtra - * @description Extra attributes to be added to input fields and their OpenAPI schema. Used by the workflow editor - * during schema parsing and UI rendering. - */ - OutputFieldJSONSchemaExtra: { - field_kind: components["schemas"]["FieldKind"]; + dither?: boolean; /** - * Ui Hidden + * Prequantize + * @description Apply 256-color quantization with specified method prior to applying the color palette * @default false */ - ui_hidden: boolean; + prequantize?: boolean; /** - * Ui Order - * @default null + * Quantizer + * @description Palettizer quantization method + * @default Fast Octree + * @enum {string} */ - ui_order: number | null; - /** @default null */ - ui_type: components["schemas"]["UIType"] | null; + quantizer?: "Median Cut" | "Max Coverage" | "Fast Octree"; + /** + * type + * @default retro_palettize + * @constant + */ + type: "retro_palettize"; }; /** - * PBR Maps - * @description Generate Normal, Displacement and Roughness Map from a given image + * Quantize + * @description Quantize an image to 256 or less colors */ - PBRMapsInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + RetroQuantizeInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -21743,87 +35695,139 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image + * @description Input image for quantizing * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Tile Size - * @description Tile size - * @default 512 + * Colors + * @description Number of colors the image should be reduced to + * @default 64 */ - tile_size?: number; + colors?: number; /** - * Border Mode - * @description Border mode to apply to eliminate any artifacts or seams - * @default none + * Method + * @description Quantization method + * @default Median Cut * @enum {string} */ - border_mode?: "none" | "seamless" | "mirror" | "replicate"; + method?: "Median Cut" | "Max Coverage" | "Fast Octree"; + /** + * Kmeans + * @description k_means + * @default 0 + */ + kmeans?: number; + /** + * Dither + * @description Dither quantized image + * @default true + */ + dither?: boolean; /** * type - * @default pbr_maps + * @default retro_quantize * @constant */ - type: "pbr_maps"; + type: "retro_quantize"; }; - /** PBRMapsOutput */ - PBRMapsOutput: { + /** + * Scan Lines + * @description Apply a simple scan lines effect to the input image + */ + RetroScanlinesSimpleInvocation: { /** - * @description The generated normal map + * @description Optional metadata to be saved with the image * @default null */ - normal_map: components["schemas"]["ImageField"]; + metadata?: components["schemas"]["MetadataField"] | null; /** - * @description The generated roughness map - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - roughness_map: components["schemas"]["ImageField"]; + id: string; /** - * @description The generated displacement map + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description Input image to add scanlines to * @default null */ - displacement_map: components["schemas"]["ImageField"]; + image?: components["schemas"]["ImageField"] | null; /** - * type - * @default pbr_maps-output - * @constant + * Line Size + * @description Thickness of scanlines in pixels + * @default 1 */ - type: "pbr_maps-output"; - }; - /** PaginatedResults[WorkflowRecordListItemWithThumbnailDTO] */ - PaginatedResults_WorkflowRecordListItemWithThumbnailDTO_: { + line_size?: number; + /** + * Line Spacing + * @description Space between lines in pixels + * @default 4 + */ + line_spacing?: number; + /** + * @description Darkness of scanlines, 1.0 being black + * @default { + * "r": 0, + * "g": 0, + * "b": 0, + * "a": 255 + * } + */ + line_color?: components["schemas"]["ColorField"]; + /** + * Size Jitter + * @description Random line position offset + * @default 0 + */ + size_jitter?: number; /** - * Page - * @description Current Page + * Space Jitter + * @description Random line position offset + * @default 0 */ - page: number; + space_jitter?: number; /** - * Pages - * @description Total number of pages + * Vertical + * @description Switch scanlines to vertical + * @default false */ - pages: number; + vertical?: boolean; /** - * Per Page - * @description Number of items per page + * type + * @default retro_scanlines_simple + * @constant */ - per_page: number; + type: "retro_scanlines_simple"; + }; + /** RetryItemsResult */ + RetryItemsResult: { /** - * Total - * @description Total number of items in result + * Queue Id + * @description The ID of the queue */ - total: number; + queue_id: string; /** - * Items - * @description Items + * Retried Item Ids + * @description The IDs of the queue items that were retried */ - items: components["schemas"]["WorkflowRecordListItemWithThumbnailDTO"][]; + retried_item_ids: number[]; }; /** - * Pair Tile with Image - * @description Pair an image with its tile properties. + * Round Float + * @description Rounds a float to a specified number of decimal places. */ - PairTileImageInvocation: { + RoundInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -21842,40 +35846,82 @@ export type components = { */ use_cache?: boolean; /** - * @description The tile image. - * @default null + * Value + * @description The float value + * @default 0 */ - image?: components["schemas"]["ImageField"] | null; + value?: number; /** - * @description The tile properties. - * @default null + * Decimals + * @description The number of decimal places + * @default 0 */ - tile?: components["schemas"]["Tile"] | null; + decimals?: number; /** * type - * @default pair_tile_image + * @default round_float * @constant */ - type: "pair_tile_image"; + type: "round_float"; }; - /** PairTileImageOutput */ - PairTileImageOutput: { - /** @description A tile description with its corresponding image. */ - tile_with_image: components["schemas"]["TileWithImage"]; + /** SAMPoint */ + SAMPoint: { + /** + * X + * @description The x-coordinate of the point + */ + x: number; + /** + * Y + * @description The y-coordinate of the point + */ + y: number; + /** @description The label of the point */ + label: components["schemas"]["SAMPointLabel"]; + }; + /** + * SAMPointLabel + * @enum {integer} + */ + SAMPointLabel: -1 | 0 | 1; + /** SAMPointsField */ + SAMPointsField: { + /** + * Points + * @description The points of the object + */ + points: components["schemas"]["SAMPoint"][]; + }; + /** + * SD3ConditioningField + * @description A conditioning tensor primitive value + */ + SD3ConditioningField: { + /** + * Conditioning Name + * @description The name of conditioning tensor + */ + conditioning_name: string; + }; + /** + * SD3ConditioningOutput + * @description Base class for nodes that output a single SD3 conditioning tensor + */ + SD3ConditioningOutput: { + /** @description Conditioning tensor */ + conditioning: components["schemas"]["SD3ConditioningField"]; /** * type - * @default pair_tile_image_output + * @default sd3_conditioning_output * @constant */ - type: "pair_tile_image_output"; + type: "sd3_conditioning_output"; }; /** - * Paste Image into Bounding Box - * @description Paste the source image into the target image at the given bounding box. - * - * The source image must be the same size as the bounding box, and the bounding box must fit within the target image. + * Denoise - SD3 + * @description Run denoising process with a SD3 model. */ - PasteImageIntoBoundingBoxInvocation: { + SD3DenoiseInvocation: { /** * @description The board to save the image to * @default null @@ -21904,32 +35950,85 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to paste + * @description Latents tensor * @default null */ - source_image?: components["schemas"]["ImageField"] | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * @description The image to paste into + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. * @default null */ - target_image?: components["schemas"]["ImageField"] | null; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * @description The bounding box to paste the image into + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 + */ + denoising_start?: number; + /** + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 + */ + denoising_end?: number; + /** + * Transformer + * @description SD3 model (MMDiTX) to load * @default null */ - bounding_box?: components["schemas"]["BoundingBoxField"] | null; + transformer?: components["schemas"]["TransformerField"] | null; + /** + * @description Positive conditioning tensor + * @default null + */ + positive_conditioning?: components["schemas"]["SD3ConditioningField"] | null; + /** + * @description Negative conditioning tensor + * @default null + */ + negative_conditioning?: components["schemas"]["SD3ConditioningField"] | null; + /** + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 3.5 + */ + cfg_scale?: number | number[]; + /** + * Width + * @description Width of the generated image. + * @default 1024 + */ + width?: number; + /** + * Height + * @description Height of the generated image. + * @default 1024 + */ + height?: number; + /** + * Steps + * @description Number of steps to run + * @default 10 + */ + steps?: number; + /** + * Seed + * @description Randomness seed for reproducibility. + * @default 0 + */ + seed?: number; /** * type - * @default paste_image_into_bounding_box + * @default sd3_denoise * @constant */ - type: "paste_image_into_bounding_box"; + type: "sd3_denoise"; }; /** - * PiDiNet Edge Detection - * @description Generates an edge map using PiDiNet. + * Image to Latents - SD3 + * @description Generates latents from an image. */ - PiDiNetEdgeDetectionInvocation: { + SD3ImageToLatentsInvocation: { /** * @description The board to save the image to * @default null @@ -21958,76 +36057,37 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The image to encode * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Quantize Edges - * @description Whether or not to use safe mode - * @default false - */ - quantize_edges?: boolean; - /** - * Scribble - * @description Whether or not to use scribble mode - * @default false + * @description VAE + * @default null */ - scribble?: boolean; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default pidi_edge_detection + * @default sd3_i2l * @constant */ - type: "pidi_edge_detection"; - }; - /** PresetData */ - PresetData: { - /** - * Positive Prompt - * @description Positive prompt - */ - positive_prompt: string; - /** - * Negative Prompt - * @description Negative prompt - */ - negative_prompt: string; + type: "sd3_i2l"; }; /** - * PresetType - * @enum {string} - */ - PresetType: "user" | "default"; - /** - * ProgressImage - * @description The progress image sent intermittently during processing + * Latents to Image - SD3 + * @description Generates an image from latents. */ - ProgressImage: { - /** - * Width - * @description The effective width of the image in pixels - */ - width: number; + SD3LatentsToImageInvocation: { /** - * Height - * @description The effective height of the image in pixels + * @description The board to save the image to + * @default null */ - height: number; + board?: components["schemas"]["BoardField"] | null; /** - * Dataurl - * @description The image data as a b64 data URL - */ - dataURL: string; - }; - /** - * Prompt Template - * @description Applies a Style Preset template to positive and negative prompts. - * - * Select a Style Preset and provide positive/negative prompts. The node replaces - * {prompt} placeholders in the template with your input prompts. - */ - PromptTemplateInvocation: { + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -22046,56 +36106,27 @@ export type components = { */ use_cache?: boolean; /** - * @description The Style Preset to use as a template + * @description Latents tensor * @default null */ - style_preset?: components["schemas"]["StylePresetField"] | null; - /** - * Positive Prompt - * @description The positive prompt to insert into the template's {prompt} placeholder - * @default - */ - positive_prompt?: string; - /** - * Negative Prompt - * @description The negative prompt to insert into the template's {prompt} placeholder - * @default - */ - negative_prompt?: string; - /** - * type - * @default prompt_template - * @constant - */ - type: "prompt_template"; - }; - /** - * PromptTemplateOutput - * @description Output for the Prompt Template node - */ - PromptTemplateOutput: { - /** - * Positive Prompt - * @description The positive prompt with the template applied - */ - positive_prompt: string; + latents?: components["schemas"]["LatentsField"] | null; /** - * Negative Prompt - * @description The negative prompt with the template applied + * @description VAE + * @default null */ - negative_prompt: string; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default prompt_template_output + * @default sd3_l2i * @constant */ - type: "prompt_template_output"; + type: "sd3_l2i"; }; /** - * Prompts from File - * @description Loads prompts from a text file + * SD3 Main Model Input + * @description Loads a sd3 model from an input, outputting its submodels. */ - PromptsFromFileInvocation: { + SD3ModelLoaderInputInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -22114,462 +36145,493 @@ export type components = { */ use_cache?: boolean; /** - * File Path - * @description Path to prompt text file + * @description SD3 model (MMDiTX) to load * @default null */ - file_path?: string | null; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Pre Prompt - * @description String to prepend to each prompt + * T5 Encoder + * @description T5 tokenizer and text encoder * @default null */ - pre_prompt?: string | null; + t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Post Prompt - * @description String to append to each prompt + * CLIP L Encoder + * @description CLIP Embed loader * @default null */ - post_prompt?: string | null; + clip_l_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Start Line - * @description Line in the file to start start from - * @default 1 + * CLIP G Encoder + * @description CLIP-G Embed loader + * @default null */ - start_line?: number; + clip_g_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Max Prompts - * @description Max lines to read from file (0=all) - * @default 1 + * VAE + * @description VAE model to load + * @default null */ - max_prompts?: number; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; /** * type - * @default prompt_from_file + * @default sd3_model_loader_input * @constant */ - type: "prompt_from_file"; + type: "sd3_model_loader_input"; }; /** - * PruneResult - * @description Result of pruning the session queue + * Prompt - SDXL + * @description Parse prompt using compel package to conditioning. */ - PruneResult: { + SDXLCompelPromptInvocation: { /** - * Deleted - * @description Number of queue items deleted + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - deleted: number; - }; - /** - * QueueClearedEvent - * @description Event model for queue_cleared - */ - QueueClearedEvent: { + id: string; /** - * Timestamp - * @description The timestamp of the event + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - timestamp: number; + is_intermediate?: boolean; /** - * Queue Id - * @description The ID of the queue + * Use Cache + * @description Whether or not to use the cache + * @default true */ - queue_id: string; - }; - /** - * QueueItemStatusChangedEvent - * @description Event model for queue_item_status_changed - */ - QueueItemStatusChangedEvent: { + use_cache?: boolean; /** - * Timestamp - * @description The timestamp of the event + * Prompt + * @description Prompt to be parsed by Compel to create a conditioning tensor + * @default */ - timestamp: number; + prompt?: string; /** - * Queue Id - * @description The ID of the queue + * Style + * @description Prompt to be parsed by Compel to create a conditioning tensor + * @default */ - queue_id: string; + style?: string; /** - * Item Id - * @description The ID of the queue item + * Original Width + * @default 1024 */ - item_id: number; + original_width?: number; /** - * Batch Id - * @description The ID of the queue batch + * Original Height + * @default 1024 */ - batch_id: string; + original_height?: number; /** - * Origin - * @description The origin of the queue item - * @default null + * Crop Top + * @default 0 */ - origin: string | null; + crop_top?: number; /** - * Destination - * @description The destination of the queue item - * @default null + * Crop Left + * @default 0 */ - destination: string | null; + crop_left?: number; /** - * User Id - * @description The ID of the user who created the queue item - * @default system + * Target Width + * @default 1024 */ - user_id: string; + target_width?: number; /** - * Status - * @description The new status of the queue item - * @enum {string} + * Target Height + * @default 1024 */ - status: "pending" | "in_progress" | "completed" | "failed" | "canceled"; + target_height?: number; /** - * Error Type - * @description The error type, if any + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - error_type: string | null; + clip?: components["schemas"]["CLIPField"] | null; /** - * Error Message - * @description The error message, if any + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - error_message: string | null; + clip2?: components["schemas"]["CLIPField"] | null; /** - * Error Traceback - * @description The error traceback, if any + * @description A mask defining the region that this conditioning prompt applies to. * @default null */ - error_traceback: string | null; + mask?: components["schemas"]["TensorField"] | null; /** - * Created At - * @description The timestamp when the queue item was created + * type + * @default sdxl_compel_prompt + * @constant */ - created_at: string; + type: "sdxl_compel_prompt"; + }; + /** + * Apply LoRA Collection - SDXL + * @description Applies a collection of SDXL LoRAs to the provided UNet and CLIP models. + */ + SDXLLoRACollectionLoader: { /** - * Updated At - * @description The timestamp when the queue item was last updated + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - updated_at: string; + id: string; /** - * Started At - * @description The timestamp when the queue item was started - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - started_at: string | null; + is_intermediate?: boolean; /** - * Completed At - * @description The timestamp when the queue item was completed - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - completed_at: string | null; - /** @description The status of the batch */ - batch_status: components["schemas"]["BatchStatus"]; - /** @description The status of the queue */ - queue_status: components["schemas"]["SessionQueueStatus"]; + use_cache?: boolean; /** - * Session Id - * @description The ID of the session (aka graph execution state) + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. + * @default null */ - session_id: string; - }; - /** - * QueueItemsRetriedEvent - * @description Event model for queue_items_retried - */ - QueueItemsRetriedEvent: { + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * Timestamp - * @description The timestamp of the event + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - timestamp: number; + unet?: components["schemas"]["UNetField"] | null; /** - * Queue Id - * @description The ID of the queue + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - queue_id: string; + clip?: components["schemas"]["CLIPField"] | null; /** - * Retried Item Ids - * @description The IDs of the queue items that were retried + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - retried_item_ids: number[]; - }; - /** - * Qwen3EncoderField - * @description Field for Qwen3 text encoder used by Z-Image models. - */ - Qwen3EncoderField: { - /** @description Info to load tokenizer submodel */ - tokenizer: components["schemas"]["ModelIdentifierField"]; - /** @description Info to load text_encoder submodel */ - text_encoder: components["schemas"]["ModelIdentifierField"]; + clip2?: components["schemas"]["CLIPField"] | null; /** - * Loras - * @description LoRAs to apply on model loading + * type + * @default sdxl_lora_collection_loader + * @constant */ - loras?: components["schemas"]["LoRAField"][]; + type: "sdxl_lora_collection_loader"; }; /** - * Qwen3Encoder_Checkpoint_Config - * @description Configuration for single-file Qwen3 Encoder models (safetensors). + * Apply LoRA - SDXL + * @description Apply selected lora to unet and text_encoder. */ - Qwen3Encoder_Checkpoint_Config: { + SDXLLoRALoaderInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * LoRA + * @description LoRA model to load + * @default null */ - file_size: number; + lora?: components["schemas"]["ModelIdentifierField"] | null; /** - * Name - * @description Name of the model. + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 */ - name: string; + weight?: number; /** - * Description - * @description Model description + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - description: string | null; + unet?: components["schemas"]["UNetField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + clip?: components["schemas"]["CLIPField"] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - source_api_response: string | null; + clip2?: components["schemas"]["CLIPField"] | null; /** - * Cover Image - * @description Url for image to preview model + * type + * @default sdxl_lora_loader + * @constant */ - cover_image: string | null; + type: "sdxl_lora_loader"; + }; + /** + * SDXLLoRALoaderOutput + * @description SDXL LoRA Loader Output + */ + SDXLLoRALoaderOutput: { /** - * Config Path - * @description Path to the config for this model, if any. + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - config_path: string | null; + unet: components["schemas"]["UNetField"] | null; /** - * Base - * @default any - * @constant + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - base: "any"; + clip: components["schemas"]["CLIPField"] | null; /** - * Type - * @default qwen3_encoder - * @constant + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - type: "qwen3_encoder"; + clip2: components["schemas"]["CLIPField"] | null; /** - * Format - * @default checkpoint + * type + * @default sdxl_lora_loader_output * @constant */ - format: "checkpoint"; - /** - * Cpu Only - * @description Whether this model should run on CPU only - */ - cpu_only: boolean | null; - /** @description Qwen3 model size variant (4B or 8B) */ - variant: components["schemas"]["Qwen3VariantType"]; + type: "sdxl_lora_loader_output"; }; /** - * Qwen3Encoder_GGUF_Config - * @description Configuration for GGUF-quantized Qwen3 Encoder models. + * SDXL Main Model Toggle + * @description Allows boolean selection between two separate SDXL Main Model inputs */ - Qwen3Encoder_GGUF_Config: { + SDXLMainModelToggleInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Use Second + * @description Use 2nd Input + * @default false */ - file_size: number; + use_second?: boolean; /** - * Name - * @description Name of the model. + * UNet 1 + * @description UNet (scheduler, LoRAs) + * @default null */ - name: string; + unet1?: components["schemas"]["UNetField"] | null; /** - * Description - * @description Model description + * CLIP 1 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - description: string | null; + clip1?: components["schemas"]["CLIPField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * CLIP 2 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + clip21?: components["schemas"]["CLIPField"] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * VAE 1 + * @description VAE + * @default null */ - source_api_response: string | null; + vae1?: components["schemas"]["VAEField"] | null; /** - * Cover Image - * @description Url for image to preview model + * UNet 2 + * @description UNet (scheduler, LoRAs) + * @default null */ - cover_image: string | null; + unet2?: components["schemas"]["UNetField"] | null; /** - * Config Path - * @description Path to the config for this model, if any. + * CLIP 1 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - config_path: string | null; + clip2?: components["schemas"]["CLIPField"] | null; /** - * Base - * @default any - * @constant + * CLIP 2 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - base: "any"; + clip22?: components["schemas"]["CLIPField"] | null; /** - * Type - * @default qwen3_encoder - * @constant + * VAE 2 + * @description VAE + * @default null */ - type: "qwen3_encoder"; + vae2?: components["schemas"]["VAEField"] | null; /** - * Format - * @default gguf_quantized + * type + * @default sdxl_main_model_toggle * @constant */ - format: "gguf_quantized"; + type: "sdxl_main_model_toggle"; + }; + /** + * SDXL Main Model Input + * @description Loads a sdxl model from an input, outputting its submodels. + */ + SDXLModelLoaderInputInvocation: { /** - * Cpu Only - * @description Whether this model should run on CPU only + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - cpu_only: boolean | null; - /** @description Qwen3 model size variant (4B or 8B) */ - variant: components["schemas"]["Qwen3VariantType"]; + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description SDXL Main model (UNet, VAE, CLIP1, CLIP2) to load + * @default null + */ + model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default sdxl_model_loader_input + * @constant + */ + type: "sdxl_model_loader_input"; }; /** - * Qwen3Encoder_Qwen3Encoder_Config - * @description Configuration for Qwen3 Encoder models in a diffusers-like format. - * - * The model weights are expected to be in a folder called text_encoder inside the model directory, - * compatible with Qwen2VLForConditionalGeneration or similar architectures used by Z-Image. + * Main Model - SDXL + * @description Loads an sdxl base model, outputting its submodels. */ - Qwen3Encoder_Qwen3Encoder_Config: { + SDXLModelLoaderInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * @description SDXL Main model (UNet, VAE, CLIP1, CLIP2) to load + * @default null */ - file_size: number; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Name - * @description Name of the model. + * type + * @default sdxl_model_loader + * @constant */ - name: string; + type: "sdxl_model_loader"; + }; + /** + * SDXLModelLoaderOutput + * @description SDXL base model loader output + */ + SDXLModelLoaderOutput: { /** - * Description - * @description Model description + * UNet + * @description UNet (scheduler, LoRAs) */ - description: string | null; + unet: components["schemas"]["UNetField"]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + clip: components["schemas"]["CLIPField"]; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - source_api_response: string | null; + clip2: components["schemas"]["CLIPField"]; /** - * Cover Image - * @description Url for image to preview model + * VAE + * @description VAE */ - cover_image: string | null; + vae: components["schemas"]["VAEField"]; /** - * Base - * @default any + * type + * @default sdxl_model_loader_output * @constant */ - base: "any"; + type: "sdxl_model_loader_output"; + }; + /** + * SDXL Model To String + * @description Converts an SDXL Model to a JSONString + */ + SDXLModelToStringInvocation: { /** - * Type - * @default qwen3_encoder - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - type: "qwen3_encoder"; + id: string; /** - * Format - * @default qwen3_encoder - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - format: "qwen3_encoder"; + is_intermediate?: boolean; /** - * Cpu Only - * @description Whether this model should run on CPU only + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cpu_only: boolean | null; - /** @description Qwen3 model size variant (4B or 8B) */ - variant: components["schemas"]["Qwen3VariantType"]; + use_cache?: boolean; + /** + * @description SDXL Main model (UNet, VAE, CLIP1, CLIP2) to load + * @default null + */ + model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default sdxl_model_to_string + * @constant + */ + type: "sdxl_model_to_string"; }; /** - * Qwen3VariantType - * @description Qwen3 text encoder variants based on model size. - * @enum {string} - */ - Qwen3VariantType: "qwen3_4b" | "qwen3_8b"; - /** - * Random Float - * @description Outputs a single random float + * Prompt - SDXL Refiner + * @description Parse prompt using compel package to conditioning. */ - RandomFloatInvocation: { + SDXLRefinerCompelPromptInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -22584,39 +36646,58 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Low - * @description The inclusive low value + * Style + * @description Prompt to be parsed by Compel to create a conditioning tensor + * @default + */ + style?: string; + /** + * Original Width + * @default 1024 + */ + original_width?: number; + /** + * Original Height + * @default 1024 + */ + original_height?: number; + /** + * Crop Top * @default 0 */ - low?: number; + crop_top?: number; /** - * High - * @description The exclusive high value - * @default 1 + * Crop Left + * @default 0 */ - high?: number; + crop_left?: number; /** - * Decimals - * @description The number of decimal places to round to - * @default 2 + * Aesthetic Score + * @description The aesthetic score to apply to the conditioning tensor + * @default 6 */ - decimals?: number; + aesthetic_score?: number; + /** + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip2?: components["schemas"]["CLIPField"] | null; /** * type - * @default rand_float + * @default sdxl_refiner_compel_prompt * @constant */ - type: "rand_float"; + type: "sdxl_refiner_compel_prompt"; }; /** - * Random Integer - * @description Outputs a single random integer. + * Refiner Model - SDXL + * @description Loads an sdxl refiner model, outputting its submodels. */ - RandomIntInvocation: { + SDXLRefinerModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -22631,33 +36712,68 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Low - * @description The inclusive low value - * @default 0 + * @description SDXL Refiner Main Modde (UNet, VAE, CLIP2) to load + * @default null */ - low?: number; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * High - * @description The exclusive high value - * @default 2147483647 + * type + * @default sdxl_refiner_model_loader + * @constant */ - high?: number; + type: "sdxl_refiner_model_loader"; + }; + /** + * SDXLRefinerModelLoaderOutput + * @description SDXL refiner model loader output + */ + SDXLRefinerModelLoaderOutput: { + /** + * UNet + * @description UNet (scheduler, LoRAs) + */ + unet: components["schemas"]["UNetField"]; + /** + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + */ + clip2: components["schemas"]["CLIPField"]; + /** + * VAE + * @description VAE + */ + vae: components["schemas"]["VAEField"]; /** * type - * @default rand_int + * @default sdxl_refiner_model_loader_output * @constant */ - type: "rand_int"; + type: "sdxl_refiner_model_loader_output"; }; /** - * Random Range - * @description Creates a collection of random numbers + * SQLiteDirection + * @enum {string} */ - RandomRangeInvocation: { + SQLiteDirection: "ASC" | "DESC"; + /** + * Save Image + * @description Saves an image. Unlike an image primitive, this invocation stores a copy of the image. + */ + SaveImageInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -22671,46 +36787,27 @@ export type components = { is_intermediate?: boolean; /** * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Low - * @description The inclusive low value - * @default 0 - */ - low?: number; - /** - * High - * @description The exclusive high value - * @default 2147483647 - */ - high?: number; - /** - * Size - * @description The number of values to generate - * @default 1 + * @description Whether or not to use the cache + * @default false */ - size?: number; + use_cache?: boolean; /** - * Seed - * @description The seed for the RNG (omit for random) - * @default 0 + * @description The image to process + * @default null */ - seed?: number; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default random_range + * @default save_image * @constant */ - type: "random_range"; + type: "save_image"; }; /** - * Integer Range - * @description Creates a range of numbers from start to stop with step + * Scale Latents + * @description Scales latents by a given factor. */ - RangeInvocation: { + ScaleLatentsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -22729,35 +36826,41 @@ export type components = { */ use_cache?: boolean; /** - * Start - * @description The start of the range - * @default 0 + * @description Latents tensor + * @default null */ - start?: number; + latents?: components["schemas"]["LatentsField"] | null; /** - * Stop - * @description The stop of the range - * @default 10 + * Scale Factor + * @description The factor by which to scale + * @default null */ - stop?: number; + scale_factor?: number | null; /** - * Step - * @description The step of the range - * @default 1 + * Mode + * @description Interpolation mode + * @default bilinear + * @enum {string} */ - step?: number; + mode?: "nearest" | "linear" | "bilinear" | "bicubic" | "trilinear" | "area" | "nearest-exact"; + /** + * Antialias + * @description Whether or not to apply antialiasing (bilinear or bicubic only) + * @default false + */ + antialias?: boolean; /** * type - * @default range + * @default lscale * @constant */ - type: "range"; + type: "lscale"; }; /** - * Integer Range of Size - * @description Creates a range from start to start + (size * step) incremented by step + * Scheduler + * @description Selects a scheduler. */ - RangeOfSizeInvocation: { + SchedulerInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -22776,194 +36879,81 @@ export type components = { */ use_cache?: boolean; /** - * Start - * @description The start of the range - * @default 0 + * Scheduler + * @description Scheduler to use during inference + * @default euler + * @enum {string} */ - start?: number; + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** - * Size - * @description The number of values - * @default 1 + * type + * @default scheduler + * @constant */ - size?: number; + type: "scheduler"; + }; + /** SchedulerOutput */ + SchedulerOutput: { /** - * Step - * @description The step of the range - * @default 1 + * Scheduler + * @description Scheduler to use during inference + * @enum {string} */ - step?: number; + scheduler: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** * type - * @default range_of_size + * @default scheduler_output * @constant */ - type: "range_of_size"; + type: "scheduler_output"; }; /** - * RecallParameter - * @description Request model for updating recallable parameters. + * SchedulerPredictionType + * @description Scheduler prediction type. + * @enum {string} */ - RecallParameter: { - /** - * Positive Prompt - * @description Positive prompt text - */ - positive_prompt?: string | null; - /** - * Negative Prompt - * @description Negative prompt text - */ - negative_prompt?: string | null; + SchedulerPredictionType: "epsilon" | "v_prediction" | "sample"; + /** + * Scheduler To String + * @description Converts a Scheduler to a string + */ + SchedulerToStringInvocation: { /** - * Model - * @description Main model name/identifier + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - model?: string | null; + id: string; /** - * Refiner Model - * @description Refiner model name/identifier + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - refiner_model?: string | null; + is_intermediate?: boolean; /** - * Vae Model - * @description VAE model name/identifier + * Use Cache + * @description Whether or not to use the cache + * @default true */ - vae_model?: string | null; + use_cache?: boolean; /** * Scheduler - * @description Scheduler name - */ - scheduler?: string | null; - /** - * Steps - * @description Number of generation steps - */ - steps?: number | null; - /** - * Refiner Steps - * @description Number of refiner steps - */ - refiner_steps?: number | null; - /** - * Cfg Scale - * @description CFG scale for guidance - */ - cfg_scale?: number | null; - /** - * Cfg Rescale Multiplier - * @description CFG rescale multiplier - */ - cfg_rescale_multiplier?: number | null; - /** - * Refiner Cfg Scale - * @description Refiner CFG scale - */ - refiner_cfg_scale?: number | null; - /** - * Guidance - * @description Guidance scale - */ - guidance?: number | null; - /** - * Width - * @description Image width in pixels - */ - width?: number | null; - /** - * Height - * @description Image height in pixels - */ - height?: number | null; - /** - * Seed - * @description Random seed - */ - seed?: number | null; - /** - * Denoise Strength - * @description Denoising strength - */ - denoise_strength?: number | null; - /** - * Refiner Denoise Start - * @description Refiner denoising start - */ - refiner_denoise_start?: number | null; - /** - * Clip Skip - * @description CLIP skip layers - */ - clip_skip?: number | null; - /** - * Seamless X - * @description Enable seamless X tiling - */ - seamless_x?: boolean | null; - /** - * Seamless Y - * @description Enable seamless Y tiling - */ - seamless_y?: boolean | null; - /** - * Refiner Positive Aesthetic Score - * @description Refiner positive aesthetic score - */ - refiner_positive_aesthetic_score?: number | null; - /** - * Refiner Negative Aesthetic Score - * @description Refiner negative aesthetic score - */ - refiner_negative_aesthetic_score?: number | null; - /** - * Loras - * @description List of LoRAs with their weights - */ - loras?: components["schemas"]["LoRARecallParameter"][] | null; - /** - * Control Layers - * @description List of control adapters (ControlNet, T2I Adapter, Control LoRA) with their settings - */ - control_layers?: components["schemas"]["ControlNetRecallParameter"][] | null; - /** - * Ip Adapters - * @description List of IP Adapters with their settings - */ - ip_adapters?: components["schemas"]["IPAdapterRecallParameter"][] | null; - }; - /** - * RecallParametersUpdatedEvent - * @description Event model for recall_parameters_updated - */ - RecallParametersUpdatedEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; - /** - * Queue Id - * @description The ID of the queue + * @description Scheduler to use during inference + * @default euler + * @enum {string} */ - queue_id: string; + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** - * Parameters - * @description The recall parameters that were updated + * type + * @default scheduler_to_string + * @constant */ - parameters: { - [key: string]: unknown; - }; + type: "scheduler_to_string"; }; /** - * Create Rectangle Mask - * @description Create a rectangular mask. + * Scheduler Toggle + * @description Allows boolean selection between two separate scheduler inputs */ - RectangleMaskInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + SchedulerToggleInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -22982,95 +36972,35 @@ export type components = { */ use_cache?: boolean; /** - * Width - * @description The width of the entire mask. - * @default null - */ - width?: number | null; - /** - * Height - * @description The height of the entire mask. - * @default null - */ - height?: number | null; - /** - * X Left - * @description The left x-coordinate of the rectangular masked region (inclusive). - * @default null - */ - x_left?: number | null; - /** - * Y Top - * @description The top y-coordinate of the rectangular masked region (inclusive). - * @default null + * Use Second + * @description Use 2nd Input + * @default false */ - y_top?: number | null; + use_second?: boolean; /** - * Rectangle Width - * @description The width of the rectangular masked region. + * Scheduler1 + * @description First Scheduler Input * @default null */ - rectangle_width?: number | null; + scheduler1?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; /** - * Rectangle Height - * @description The height of the rectangular masked region. + * Scheduler2 + * @description Second Scheduler Input * @default null */ - rectangle_height?: number | null; + scheduler2?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; /** * type - * @default rectangle_mask + * @default scheduler_toggle * @constant */ - type: "rectangle_mask"; - }; - /** - * RemoteModelFile - * @description Information about a downloadable file that forms part of a model. - */ - RemoteModelFile: { - /** - * Url - * Format: uri - * @description The url to download this model file - */ - url: string; - /** - * Path - * Format: path - * @description The path to the file, relative to the model root - */ - path: string; - /** - * Size - * @description The size of this file, in bytes - * @default 0 - */ - size?: number | null; - /** - * Sha256 - * @description SHA256 hash of this model (not always available) - */ - sha256?: string | null; - }; - /** RemoveImagesFromBoardResult */ - RemoveImagesFromBoardResult: { - /** - * Affected Boards - * @description The ids of boards affected by the delete operation - */ - affected_boards: string[]; - /** - * Removed Images - * @description The image names that were removed from their board - */ - removed_images: string[]; + type: "scheduler_toggle"; }; /** - * Resize Latents - * @description Resizes latents to explicit width/height (in pixels). Provided dimensions are floor-divided by 8. + * Main Model - SD3 + * @description Loads a SD3 base model, outputting its submodels. */ - ResizeLatentsInvocation: { + Sd3ModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23088,71 +37018,81 @@ export type components = { * @default true */ use_cache?: boolean; + /** @description SD3 model (MMDiTX) to load */ + model: components["schemas"]["ModelIdentifierField"]; /** - * @description Latents tensor + * T5 Encoder + * @description T5 tokenizer and text encoder * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Width - * @description Width of output (px) + * CLIP L Encoder + * @description CLIP Embed loader * @default null */ - width?: number | null; + clip_l_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Height - * @description Width of output (px) + * CLIP G Encoder + * @description CLIP-G Embed loader * @default null */ - height?: number | null; - /** - * Mode - * @description Interpolation mode - * @default bilinear - * @enum {string} - */ - mode?: "nearest" | "linear" | "bilinear" | "bicubic" | "trilinear" | "area" | "nearest-exact"; + clip_g_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Antialias - * @description Whether or not to apply antialiasing (bilinear or bicubic only) - * @default false + * VAE + * @description VAE model to load + * @default null */ - antialias?: boolean; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; /** * type - * @default lresize + * @default sd3_model_loader * @constant */ - type: "lresize"; + type: "sd3_model_loader"; }; /** - * ResourceOrigin - * @description The origin of a resource (eg image). - * - * - INTERNAL: The resource was created by the application. - * - EXTERNAL: The resource was not created by the application. - * This may be a user-initiated upload, or an internal application upload (eg Canvas init image). - * @enum {string} + * Sd3ModelLoaderOutput + * @description SD3 base model loader output. */ - ResourceOrigin: "internal" | "external"; - /** RetryItemsResult */ - RetryItemsResult: { + Sd3ModelLoaderOutput: { /** - * Queue Id - * @description The ID of the queue + * Transformer + * @description Transformer */ - queue_id: string; + transformer: components["schemas"]["TransformerField"]; /** - * Retried Item Ids - * @description The IDs of the queue items that were retried + * CLIP L + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - retried_item_ids: number[]; + clip_l: components["schemas"]["CLIPField"]; + /** + * CLIP G + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + */ + clip_g: components["schemas"]["CLIPField"]; + /** + * T5 Encoder + * @description T5 tokenizer and text encoder + */ + t5_encoder: components["schemas"]["T5EncoderField"]; + /** + * VAE + * @description VAE + */ + vae: components["schemas"]["VAEField"]; + /** + * type + * @default sd3_model_loader_output + * @constant + */ + type: "sd3_model_loader_output"; }; /** - * Round Float - * @description Rounds a float to a specified number of decimal places. + * SD3 Model To String + * @description Converts an SD3 Model to a JSONString */ - RoundInvocation: { + Sd3ModelToStringInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23171,92 +37111,22 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The float value - * @default 0 - */ - value?: number; - /** - * Decimals - * @description The number of decimal places - * @default 0 - */ - decimals?: number; - /** - * type - * @default round_float - * @constant - */ - type: "round_float"; - }; - /** SAMPoint */ - SAMPoint: { - /** - * X - * @description The x-coordinate of the point - */ - x: number; - /** - * Y - * @description The y-coordinate of the point - */ - y: number; - /** @description The label of the point */ - label: components["schemas"]["SAMPointLabel"]; - }; - /** - * SAMPointLabel - * @enum {integer} - */ - SAMPointLabel: -1 | 0 | 1; - /** SAMPointsField */ - SAMPointsField: { - /** - * Points - * @description The points of the object - */ - points: components["schemas"]["SAMPoint"][]; - }; - /** - * SD3ConditioningField - * @description A conditioning tensor primitive value - */ - SD3ConditioningField: { - /** - * Conditioning Name - * @description The name of conditioning tensor + * @description SD3 model (MMDiTX) to load + * @default null */ - conditioning_name: string; - }; - /** - * SD3ConditioningOutput - * @description Base class for nodes that output a single SD3 conditioning tensor - */ - SD3ConditioningOutput: { - /** @description Conditioning tensor */ - conditioning: components["schemas"]["SD3ConditioningField"]; + model?: components["schemas"]["ModelIdentifierField"] | null; /** * type - * @default sd3_conditioning_output + * @default sd3_model_to_string * @constant */ - type: "sd3_conditioning_output"; + type: "sd3_model_to_string"; }; /** - * Denoise - SD3 - * @description Run denoising process with a SD3 model. - */ - SD3DenoiseInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + * Prompt - SD3 + * @description Encodes and preps a prompt for a SD3 image. + */ + Sd3TextEncoderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23275,95 +37145,41 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * CLIP L + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; - /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 - */ - denoising_start?: number; - /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 - */ - denoising_end?: number; + clip_l?: components["schemas"]["CLIPField"] | null; /** - * Transformer - * @description SD3 model (MMDiTX) to load + * CLIP G + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - transformer?: components["schemas"]["TransformerField"] | null; + clip_g?: components["schemas"]["CLIPField"] | null; /** - * @description Positive conditioning tensor + * T5Encoder + * @description T5 tokenizer and text encoder * @default null */ - positive_conditioning?: components["schemas"]["SD3ConditioningField"] | null; + t5_encoder?: components["schemas"]["T5EncoderField"] | null; /** - * @description Negative conditioning tensor + * Prompt + * @description Text prompt to encode. * @default null */ - negative_conditioning?: components["schemas"]["SD3ConditioningField"] | null; - /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 3.5 - */ - cfg_scale?: number | number[]; - /** - * Width - * @description Width of the generated image. - * @default 1024 - */ - width?: number; - /** - * Height - * @description Height of the generated image. - * @default 1024 - */ - height?: number; - /** - * Steps - * @description Number of steps to run - * @default 10 - */ - steps?: number; - /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 - */ - seed?: number; + prompt?: string | null; /** * type - * @default sd3_denoise + * @default sd3_text_encoder * @constant */ - type: "sd3_denoise"; + type: "sd3_text_encoder"; }; /** - * Image to Latents - SD3 - * @description Generates latents from an image. + * Apply Seamless - SD1.5, SDXL + * @description Applies the seamless transformation to the Model UNet and VAE. */ - SD3ImageToLatentsInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + SeamlessModeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23382,37 +37198,65 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to encode + * UNet + * @description UNet (scheduler, LoRAs) * @default null */ - image?: components["schemas"]["ImageField"] | null; + unet?: components["schemas"]["UNetField"] | null; /** - * @description VAE + * VAE + * @description VAE model to load * @default null */ vae?: components["schemas"]["VAEField"] | null; + /** + * Seamless Y + * @description Specify whether Y axis is seamless + * @default true + */ + seamless_y?: boolean; + /** + * Seamless X + * @description Specify whether X axis is seamless + * @default true + */ + seamless_x?: boolean; /** * type - * @default sd3_i2l + * @default seamless * @constant */ - type: "sd3_i2l"; + type: "seamless"; }; /** - * Latents to Image - SD3 - * @description Generates an image from latents. + * SeamlessModeOutput + * @description Modified Seamless Model output */ - SD3LatentsToImageInvocation: { + SeamlessModeOutput: { /** - * @description The board to save the image to + * UNet + * @description UNet (scheduler, LoRAs) * @default null */ - board?: components["schemas"]["BoardField"] | null; + unet: components["schemas"]["UNetField"] | null; /** - * @description Optional metadata to be saved with the image + * VAE + * @description VAE * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + vae: components["schemas"]["VAEField"] | null; + /** + * type + * @default seamless_output + * @constant + */ + type: "seamless_output"; + }; + /** + * Segment Anything + * @description Runs a Segment Anything Model (SAM or SAM2). + */ + SegmentAnythingInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23431,27 +37275,54 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * Model + * @description The Segment Anything model to use (SAM or SAM2). * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + model?: ("segment-anything-base" | "segment-anything-large" | "segment-anything-huge" | "segment-anything-2-tiny" | "segment-anything-2-small" | "segment-anything-2-base" | "segment-anything-2-large") | null; /** - * @description VAE + * @description The image to segment. * @default null */ - vae?: components["schemas"]["VAEField"] | null; + image?: components["schemas"]["ImageField"] | null; + /** + * Bounding Boxes + * @description The bounding boxes to prompt the model with. + * @default null + */ + bounding_boxes?: components["schemas"]["BoundingBoxField"][] | null; + /** + * Point Lists + * @description The list of point lists to prompt the model with. Each list of points represents a single object. + * @default null + */ + point_lists?: components["schemas"]["SAMPointsField"][] | null; + /** + * Apply Polygon Refinement + * @description Whether to apply polygon refinement to the masks. This will smooth the edges of the masks slightly and ensure that each mask consists of a single closed polygon (before merging). + * @default true + */ + apply_polygon_refinement?: boolean; + /** + * Mask Filter + * @description The filtering to apply to the detected masks before merging them into a final output. + * @default all + * @enum {string} + */ + mask_filter?: "all" | "largest" | "highest_box_score"; /** * type - * @default sd3_l2i + * @default segment_anything * @constant */ - type: "sd3_l2i"; + type: "segment_anything"; }; /** - * Prompt - SDXL - * @description Parse prompt using compel package to conditioning. + * Separate Prompt and Seed Vector + * @description Parses a JSON string representing a list of two strings, + * outputting each string separately. */ - SDXLCompelPromptInvocation: { + SeparatePromptAndSeedVectorInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23470,284 +37341,331 @@ export type components = { */ use_cache?: boolean; /** - * Prompt - * @description Prompt to be parsed by Compel to create a conditioning tensor - * @default + * Pair Input + * @description JSON string of a list containing exactly two strings, e.g., '["string one", "string two"]' + * @default ["", ""] */ - prompt?: string; + pair_input?: string; /** - * Style - * @description Prompt to be parsed by Compel to create a conditioning tensor - * @default + * type + * @default separate_prompt_and_seed_vector + * @constant */ - style?: string; + type: "separate_prompt_and_seed_vector"; + }; + /** SessionProcessorStatus */ + SessionProcessorStatus: { /** - * Original Width - * @default 1024 + * Is Started + * @description Whether the session processor is started + */ + is_started: boolean; + /** + * Is Processing + * @description Whether a session is being processed + */ + is_processing: boolean; + }; + /** + * SessionQueueAndProcessorStatus + * @description The overall status of session queue and processor + */ + SessionQueueAndProcessorStatus: { + queue: components["schemas"]["SessionQueueStatus"]; + processor: components["schemas"]["SessionProcessorStatus"]; + }; + /** SessionQueueCountsByDestination */ + SessionQueueCountsByDestination: { + /** + * Queue Id + * @description The ID of the queue + */ + queue_id: string; + /** + * Destination + * @description The destination of queue items included in this status + */ + destination: string; + /** + * Pending + * @description Number of queue items with status 'pending' for the destination + */ + pending: number; + /** + * In Progress + * @description Number of queue items with status 'in_progress' for the destination + */ + in_progress: number; + /** + * Completed + * @description Number of queue items with status 'complete' for the destination + */ + completed: number; + /** + * Failed + * @description Number of queue items with status 'error' for the destination + */ + failed: number; + /** + * Canceled + * @description Number of queue items with status 'canceled' for the destination + */ + canceled: number; + /** + * Total + * @description Total number of queue items for the destination */ - original_width?: number; + total: number; + }; + /** + * SessionQueueItem + * @description Session queue item without the full graph. Used for serialization. + */ + SessionQueueItem: { /** - * Original Height - * @default 1024 + * Item Id + * @description The identifier of the session queue item */ - original_height?: number; + item_id: number; /** - * Crop Top - * @default 0 + * Status + * @description The status of this queue item + * @default pending + * @enum {string} */ - crop_top?: number; + status: "pending" | "in_progress" | "completed" | "failed" | "canceled"; /** - * Crop Left + * Priority + * @description The priority of this queue item * @default 0 */ - crop_left?: number; + priority: number; /** - * Target Width - * @default 1024 + * Batch Id + * @description The ID of the batch associated with this queue item */ - target_width?: number; + batch_id: string; /** - * Target Height - * @default 1024 + * Origin + * @description The origin of this queue item. This data is used by the frontend to determine how to handle results. */ - target_height?: number; + origin?: string | null; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Destination + * @description The origin of this queue item. This data is used by the frontend to determine how to handle results */ - clip?: components["schemas"]["CLIPField"] | null; + destination?: string | null; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Session Id + * @description The ID of the session associated with this queue item. The session doesn't exist in graph_executions until the queue item is executed. */ - clip2?: components["schemas"]["CLIPField"] | null; + session_id: string; /** - * @description A mask defining the region that this conditioning prompt applies to. - * @default null + * Error Type + * @description The error type if this queue item errored */ - mask?: components["schemas"]["TensorField"] | null; + error_type?: string | null; /** - * type - * @default sdxl_compel_prompt - * @constant + * Error Message + * @description The error message if this queue item errored */ - type: "sdxl_compel_prompt"; - }; - /** - * Apply LoRA Collection - SDXL - * @description Applies a collection of SDXL LoRAs to the provided UNet and CLIP models. - */ - SDXLLoRACollectionLoader: { + error_message?: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Error Traceback + * @description The error traceback if this queue item errored */ - id: string; + error_traceback?: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Created At + * @description When this queue item was created */ - is_intermediate?: boolean; + created_at: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Updated At + * @description When this queue item was updated */ - use_cache?: boolean; + updated_at: string; /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. - * @default null + * Started At + * @description When this queue item was started */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; + started_at?: string | null; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Completed At + * @description When this queue item was completed */ - unet?: components["schemas"]["UNetField"] | null; + completed_at?: string | null; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Queue Id + * @description The id of the queue with which this item is associated */ - clip?: components["schemas"]["CLIPField"] | null; + queue_id: string; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * User Id + * @description The id of the user who created this queue item + * @default system */ - clip2?: components["schemas"]["CLIPField"] | null; + user_id?: string; /** - * type - * @default sdxl_lora_collection_loader - * @constant + * User Display Name + * @description The display name of the user who created this queue item, if available */ - type: "sdxl_lora_collection_loader"; - }; - /** - * Apply LoRA - SDXL - * @description Apply selected lora to unet and text_encoder. - */ - SDXLLoRALoaderInvocation: { + user_display_name?: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * User Email + * @description The email of the user who created this queue item, if available */ - id: string; + user_email?: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Field Values + * @description The field values that were used for this queue item */ - is_intermediate?: boolean; + field_values?: components["schemas"]["NodeFieldValue"][] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Retried From Item Id + * @description The item_id of the queue item that this item was retried from */ - use_cache?: boolean; + retried_from_item_id?: number | null; + /** @description The fully-populated session to be executed */ + session: components["schemas"]["GraphExecutionState"]; + /** @description The workflow associated with this queue item */ + workflow?: components["schemas"]["WorkflowWithoutID"] | null; + }; + /** SessionQueueStatus */ + SessionQueueStatus: { /** - * LoRA - * @description LoRA model to load - * @default null + * Queue Id + * @description The ID of the queue */ - lora?: components["schemas"]["ModelIdentifierField"] | null; + queue_id: string; /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 + * Item Id + * @description The current queue item id */ - weight?: number; + item_id: number | null; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Batch Id + * @description The current queue item's batch id */ - unet?: components["schemas"]["UNetField"] | null; + batch_id: string | null; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Session Id + * @description The current queue item's session id */ - clip?: components["schemas"]["CLIPField"] | null; + session_id: string | null; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Pending + * @description Number of queue items with status 'pending' */ - clip2?: components["schemas"]["CLIPField"] | null; + pending: number; /** - * type - * @default sdxl_lora_loader - * @constant + * In Progress + * @description Number of queue items with status 'in_progress' */ - type: "sdxl_lora_loader"; - }; - /** - * SDXLLoRALoaderOutput - * @description SDXL LoRA Loader Output - */ - SDXLLoRALoaderOutput: { + in_progress: number; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Completed + * @description Number of queue items with status 'complete' */ - unet: components["schemas"]["UNetField"] | null; + completed: number; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Failed + * @description Number of queue items with status 'error' */ - clip: components["schemas"]["CLIPField"] | null; + failed: number; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Canceled + * @description Number of queue items with status 'canceled' */ - clip2: components["schemas"]["CLIPField"] | null; + canceled: number; /** - * type - * @default sdxl_lora_loader_output - * @constant + * Total + * @description Total number of queue items */ - type: "sdxl_lora_loader_output"; - }; - /** - * Main Model - SDXL - * @description Loads an sdxl base model, outputting its submodels. - */ - SDXLModelLoaderInvocation: { + total: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * User Pending + * @description Number of queue items with status 'pending' for the current user */ - id: string; + user_pending?: number | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * User In Progress + * @description Number of queue items with status 'in_progress' for the current user */ - is_intermediate?: boolean; + user_in_progress?: number | null; + }; + /** + * SetupRequest + * @description Request body for initial admin setup. + */ + SetupRequest: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Email + * @description Admin email address */ - use_cache?: boolean; + email: string; /** - * @description SDXL Main model (UNet, VAE, CLIP1, CLIP2) to load - * @default null + * Display Name + * @description Admin display name */ - model?: components["schemas"]["ModelIdentifierField"] | null; + display_name?: string | null; /** - * type - * @default sdxl_model_loader - * @constant + * Password + * @description Admin password */ - type: "sdxl_model_loader"; + password: string; }; /** - * SDXLModelLoaderOutput - * @description SDXL base model loader output + * SetupResponse + * @description Response from successful admin setup. */ - SDXLModelLoaderOutput: { - /** - * UNet - * @description UNet (scheduler, LoRAs) - */ - unet: components["schemas"]["UNetField"]; + SetupResponse: { /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Success + * @description Whether setup was successful */ - clip: components["schemas"]["CLIPField"]; + success: boolean; + /** @description Created admin user information */ + user: components["schemas"]["UserDTO"]; + }; + /** + * SetupStatusResponse + * @description Response for setup status check. + */ + SetupStatusResponse: { /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Setup Required + * @description Whether initial setup is required */ - clip2: components["schemas"]["CLIPField"]; + setup_required: boolean; /** - * VAE - * @description VAE + * Multiuser Enabled + * @description Whether multiuser mode is enabled */ - vae: components["schemas"]["VAEField"]; + multiuser_enabled: boolean; /** - * type - * @default sdxl_model_loader_output - * @constant + * Strict Password Checking + * @description Whether strict password requirements are enforced */ - type: "sdxl_model_loader_output"; + strict_password_checking: boolean; }; /** - * Prompt - SDXL Refiner - * @description Parse prompt using compel package to conditioning. + * Shadows/Highlights/Midtones + * @description Extract a Shadows/Highlights/Midtones mask from an image. + * + * Extract three masks (with adjustable hard or soft thresholds) representing shadows, midtones, and highlights regions of an image. */ - SDXLRefinerCompelPromptInvocation: { + ShadowsHighlightsMidtonesMaskInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23766,54 +37684,89 @@ export type components = { */ use_cache?: boolean; /** - * Style - * @description Prompt to be parsed by Compel to create a conditioning tensor - * @default + * @description Image from which to extract mask + * @default null */ - style?: string; + image?: components["schemas"]["ImageField"] | null; /** - * Original Width - * @default 1024 + * Invert Output + * @description Off: white on black / On: black on white + * @default true */ - original_width?: number; + invert_output?: boolean; /** - * Original Height - * @default 1024 + * Highlight Threshold + * @description Threshold beyond which mask values will be at extremum + * @default 0.75 */ - original_height?: number; + highlight_threshold?: number; /** - * Crop Top + * Upper Mid Threshold + * @description Threshold to which to extend mask border by 0..1 gradient + * @default 0.7 + */ + upper_mid_threshold?: number; + /** + * Lower Mid Threshold + * @description Threshold to which to extend mask border by 0..1 gradient + * @default 0.3 + */ + lower_mid_threshold?: number; + /** + * Shadow Threshold + * @description Threshold beyond which mask values will be at extremum + * @default 0.25 + */ + shadow_threshold?: number; + /** + * Mask Expand Or Contract + * @description Pixels to grow (or shrink) the mask areas * @default 0 */ - crop_top?: number; + mask_expand_or_contract?: number; /** - * Crop Left + * Mask Blur + * @description Gaussian blur radius to apply to the masks * @default 0 */ - crop_left?: number; + mask_blur?: number; /** - * Aesthetic Score - * @description The aesthetic score to apply to the conditioning tensor - * @default 6 + * type + * @default shmmask + * @constant */ - aesthetic_score?: number; + type: "shmmask"; + }; + /** ShadowsHighlightsMidtonesMasksOutput */ + ShadowsHighlightsMidtonesMasksOutput: { + /** @description Soft-edged highlights mask */ + highlights_mask: components["schemas"]["ImageField"]; + /** @description Soft-edged midtones mask */ + midtones_mask: components["schemas"]["ImageField"]; + /** @description Soft-edged shadows mask */ + shadows_mask: components["schemas"]["ImageField"]; /** - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Width + * @description Width of the input/outputs */ - clip2?: components["schemas"]["CLIPField"] | null; + width: number; + /** + * Height + * @description Height of the input/outputs + */ + height: number; /** * type - * @default sdxl_refiner_compel_prompt + * @default shmmask_output * @constant */ - type: "sdxl_refiner_compel_prompt"; + type: "shmmask_output"; }; /** - * Refiner Model - SDXL - * @description Loads an sdxl refiner model, outputting its submodels. + * Show Image + * @description Displays a provided image using the OS image viewer, and passes it forward in the pipeline. */ - SDXLRefinerModelLoaderInvocation: { + ShowImageInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23832,98 +37785,110 @@ export type components = { */ use_cache?: boolean; /** - * @description SDXL Refiner Main Modde (UNet, VAE, CLIP2) to load + * @description The image to show * @default null */ - model?: components["schemas"]["ModelIdentifierField"] | null; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default sdxl_refiner_model_loader + * @default show_image * @constant */ - type: "sdxl_refiner_model_loader"; + type: "show_image"; }; /** - * SDXLRefinerModelLoaderOutput - * @description SDXL refiner model loader output + * SigLIP_Diffusers_Config + * @description Model config for SigLIP. */ - SDXLRefinerModelLoaderOutput: { + SigLIP_Diffusers_Config: { /** - * UNet - * @description UNet (scheduler, LoRAs) + * Key + * @description A unique key for this model. */ - unet: components["schemas"]["UNetField"]; + key: string; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Hash + * @description The hash of the model file(s). */ - clip2: components["schemas"]["CLIPField"]; + hash: string; /** - * VAE - * @description VAE + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - vae: components["schemas"]["VAEField"]; + path: string; /** - * type - * @default sdxl_refiner_model_loader_output - * @constant + * File Size + * @description The size of the model in bytes. */ - type: "sdxl_refiner_model_loader_output"; - }; - /** - * SQLiteDirection - * @enum {string} - */ - SQLiteDirection: "ASC" | "DESC"; - /** - * Save Image - * @description Saves an image. Unlike an image primitive, this invocation stores a copy of the image. - */ - SaveImageInvocation: { + file_size: number; /** - * @description The board to save the image to - * @default null + * Name + * @description Name of the model. */ - board?: components["schemas"]["BoardField"] | null; + name: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Description + * @description Model description */ - metadata?: components["schemas"]["MetadataField"] | null; + description: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source + * @description The original source of the model (path, URL or repo_id). */ - id: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - is_intermediate?: boolean; + source_api_response: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Cover Image + * @description Url for image to preview model */ - use_cache?: boolean; + cover_image: string | null; /** - * @description The image to process + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + /** + * Type + * @default siglip + * @constant + */ + type: "siglip"; + /** + * Base + * @default any + * @constant + */ + base: "any"; + /** + * Cpu Only + * @description Whether this model should run on CPU only + */ + cpu_only: boolean | null; + }; + /** + * Image-to-Image (Autoscale) + * @description Run any spandrel image-to-image model (https://github.com/chaiNNer-org/spandrel) until the target scale is reached. + */ + SpandrelImageToImageAutoscaleInvocation: { + /** + * @description The board to save the image to * @default null */ - image?: components["schemas"]["ImageField"] | null; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default save_image - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "save_image"; - }; - /** - * Scale Latents - * @description Scales latents by a given factor. - */ - ScaleLatentsInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23942,41 +37907,56 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * @description The input image * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Scale Factor - * @description The factor by which to scale + * Image-to-Image Model + * @description Image-to-Image model * @default null */ - scale_factor?: number | null; - /** - * Mode - * @description Interpolation mode - * @default bilinear - * @enum {string} - */ - mode?: "nearest" | "linear" | "bilinear" | "bicubic" | "trilinear" | "area" | "nearest-exact"; + image_to_image_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Antialias - * @description Whether or not to apply antialiasing (bilinear or bicubic only) - * @default false + * Tile Size + * @description The tile size for tiled image-to-image. Set to 0 to disable tiling. + * @default 512 */ - antialias?: boolean; + tile_size?: number; /** * type - * @default lscale + * @default spandrel_image_to_image_autoscale * @constant */ - type: "lscale"; + type: "spandrel_image_to_image_autoscale"; + /** + * Scale + * @description The final scale of the output image. If the model does not upscale the image, this will be ignored. + * @default 4 + */ + scale?: number; + /** + * Fit To Multiple Of 8 + * @description If true, the output image will be resized to the nearest multiple of 8 in both dimensions. + * @default false + */ + fit_to_multiple_of_8?: boolean; }; /** - * Scheduler - * @description Selects a scheduler. + * Image-to-Image + * @description Run any spandrel image-to-image model (https://github.com/chaiNNer-org/spandrel). */ - SchedulerInvocation: { + SpandrelImageToImageInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -23995,137 +37975,115 @@ export type components = { */ use_cache?: boolean; /** - * Scheduler - * @description Scheduler to use during inference - * @default euler - * @enum {string} + * @description The input image + * @default null */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + image?: components["schemas"]["ImageField"] | null; /** - * type - * @default scheduler - * @constant + * Image-to-Image Model + * @description Image-to-Image model + * @default null */ - type: "scheduler"; - }; - /** SchedulerOutput */ - SchedulerOutput: { + image_to_image_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Scheduler - * @description Scheduler to use during inference - * @enum {string} + * Tile Size + * @description The tile size for tiled image-to-image. Set to 0 to disable tiling. + * @default 512 */ - scheduler: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + tile_size?: number; /** * type - * @default scheduler_output + * @default spandrel_image_to_image * @constant */ - type: "scheduler_output"; + type: "spandrel_image_to_image"; }; /** - * SchedulerPredictionType - * @description Scheduler prediction type. - * @enum {string} - */ - SchedulerPredictionType: "epsilon" | "v_prediction" | "sample"; - /** - * Main Model - SD3 - * @description Loads a SD3 base model, outputting its submodels. + * Spandrel_Checkpoint_Config + * @description Model config for Spandrel Image to Image models. */ - Sd3ModelLoaderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + Spandrel_Checkpoint_Config: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Key + * @description A unique key for this model. */ - use_cache?: boolean; - /** @description SD3 model (MMDiTX) to load */ - model: components["schemas"]["ModelIdentifierField"]; + key: string; /** - * T5 Encoder - * @description T5 tokenizer and text encoder - * @default null + * Hash + * @description The hash of the model file(s). */ - t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; + hash: string; /** - * CLIP L Encoder - * @description CLIP Embed loader - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - clip_l_model?: components["schemas"]["ModelIdentifierField"] | null; + path: string; /** - * CLIP G Encoder - * @description CLIP-G Embed loader - * @default null + * File Size + * @description The size of the model in bytes. */ - clip_g_model?: components["schemas"]["ModelIdentifierField"] | null; + file_size: number; /** - * VAE - * @description VAE model to load - * @default null + * Name + * @description Name of the model. */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + name: string; /** - * type - * @default sd3_model_loader - * @constant + * Description + * @description Model description */ - type: "sd3_model_loader"; - }; - /** - * Sd3ModelLoaderOutput - * @description SD3 base model loader output. - */ - Sd3ModelLoaderOutput: { + description: string | null; /** - * Transformer - * @description Transformer + * Source + * @description The original source of the model (path, URL or repo_id). */ - transformer: components["schemas"]["TransformerField"]; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * CLIP L - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - clip_l: components["schemas"]["CLIPField"]; + source_api_response: string | null; /** - * CLIP G - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Cover Image + * @description Url for image to preview model */ - clip_g: components["schemas"]["CLIPField"]; + cover_image: string | null; /** - * T5 Encoder - * @description T5 tokenizer and text encoder + * Base + * @default any + * @constant */ - t5_encoder: components["schemas"]["T5EncoderField"]; + base: "any"; /** - * VAE - * @description VAE + * Type + * @default spandrel_image_to_image + * @constant */ - vae: components["schemas"]["VAEField"]; + type: "spandrel_image_to_image"; /** - * type - * @default sd3_model_loader_output + * Format + * @default checkpoint * @constant */ - type: "sd3_model_loader_output"; + format: "checkpoint"; }; /** - * Prompt - SD3 - * @description Encodes and preps a prompt for a SD3 image. + * Spherical Distortion + * @description Applies spherical distortion to an image and fills the frame with the resulting image */ - Sd3TextEncoderInvocation: { + SphericalDistortionInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -24144,41 +38102,123 @@ export type components = { */ use_cache?: boolean; /** - * CLIP L - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * @description The image to distort + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * K1 + * @description k1 + * @default 0.3 + */ + k1?: number; + /** + * K2 + * @description k2 + * @default 0.1 + */ + k2?: number; + /** + * P1 + * @description p1 + * @default 0 + */ + p1?: number; + /** + * P2 + * @description p2 + * @default 0 + */ + p2?: number; + /** + * type + * @default spherical_distortion + * @constant + */ + type: "spherical_distortion"; + }; + /** StarredImagesResult */ + StarredImagesResult: { + /** + * Affected Boards + * @description The ids of boards affected by the delete operation + */ + affected_boards: string[]; + /** + * Starred Images + * @description The names of the images that were starred */ - clip_l?: components["schemas"]["CLIPField"] | null; + starred_images: string[]; + }; + /** StarterModel */ + StarterModel: { + /** Description */ + description: string; + /** Source */ + source: string; + /** Name */ + name: string; + base: components["schemas"]["BaseModelType"]; + type: components["schemas"]["ModelType"]; + format?: components["schemas"]["ModelFormat"] | null; /** - * CLIP G - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Is Installed + * @default false */ - clip_g?: components["schemas"]["CLIPField"] | null; + is_installed?: boolean; /** - * T5Encoder - * @description T5 tokenizer and text encoder - * @default null + * Previous Names + * @default [] */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; + previous_names?: string[]; + /** Dependencies */ + dependencies?: components["schemas"]["StarterModelWithoutDependencies"][] | null; + }; + /** StarterModelBundle */ + StarterModelBundle: { + /** Name */ + name: string; + /** Models */ + models: components["schemas"]["StarterModel"][]; + }; + /** StarterModelResponse */ + StarterModelResponse: { + /** Starter Models */ + starter_models: components["schemas"]["StarterModel"][]; + /** Starter Bundles */ + starter_bundles: { + [key: string]: components["schemas"]["StarterModelBundle"]; + }; + }; + /** StarterModelWithoutDependencies */ + StarterModelWithoutDependencies: { + /** Description */ + description: string; + /** Source */ + source: string; + /** Name */ + name: string; + base: components["schemas"]["BaseModelType"]; + type: components["schemas"]["ModelType"]; + format?: components["schemas"]["ModelFormat"] | null; /** - * Prompt - * @description Text prompt to encode. - * @default null + * Is Installed + * @default false */ - prompt?: string | null; + is_installed?: boolean; /** - * type - * @default sd3_text_encoder - * @constant + * Previous Names + * @default [] */ - type: "sd3_text_encoder"; + previous_names?: string[]; }; /** - * Apply Seamless - SD1.5, SDXL - * @description Applies the seamless transformation to the Model UNet and VAE. + * Store Flux Conditioning + * @description Stores a FLUX Conditioning object (CLIP and T5 embeddings) into an SQLite database. + * Returns a unique identifier for retrieval. + * Includes database size management with proactive deletion and VACUUM. */ - SeamlessModeInvocation: { + StoreFluxConditioningInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -24193,69 +38233,48 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; - /** - * VAE - * @description VAE model to load + * @description The FLUX Conditioning object to store. * @default null */ - vae?: components["schemas"]["VAEField"] | null; - /** - * Seamless Y - * @description Specify whether Y axis is seamless - * @default true - */ - seamless_y?: boolean; - /** - * Seamless X - * @description Specify whether X axis is seamless - * @default true - */ - seamless_x?: boolean; + conditioning?: components["schemas"]["FluxConditioningField"] | null; /** * type - * @default seamless + * @default store_flux_conditioning * @constant */ - type: "seamless"; + type: "store_flux_conditioning"; }; /** - * SeamlessModeOutput - * @description Modified Seamless Model output + * String2Output + * @description Base class for invocations that output two strings */ - SeamlessModeOutput: { + String2Output: { /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * String 1 + * @description string 1 */ - unet: components["schemas"]["UNetField"] | null; + string_1: string; /** - * VAE - * @description VAE - * @default null + * String 2 + * @description string 2 */ - vae: components["schemas"]["VAEField"] | null; + string_2: string; /** * type - * @default seamless_output + * @default string_2_output * @constant */ - type: "seamless_output"; + type: "string_2_output"; }; /** - * Segment Anything - * @description Runs a Segment Anything Model (SAM or SAM2). + * String Batch + * @description Create a batched generation, where the workflow is executed once for each string in the batch. */ - SegmentAnythingInvocation: { + StringBatchInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -24274,349 +38293,334 @@ export type components = { */ use_cache?: boolean; /** - * Model - * @description The Segment Anything model to use (SAM or SAM2). - * @default null - */ - model?: ("segment-anything-base" | "segment-anything-large" | "segment-anything-huge" | "segment-anything-2-tiny" | "segment-anything-2-small" | "segment-anything-2-base" | "segment-anything-2-large") | null; - /** - * @description The image to segment. - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Bounding Boxes - * @description The bounding boxes to prompt the model with. - * @default null + * Batch Group + * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. + * @default None + * @enum {string} */ - bounding_boxes?: components["schemas"]["BoundingBoxField"][] | null; + batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; /** - * Point Lists - * @description The list of point lists to prompt the model with. Each list of points represents a single object. + * Strings + * @description The strings to batch over * @default null */ - point_lists?: components["schemas"]["SAMPointsField"][] | null; - /** - * Apply Polygon Refinement - * @description Whether to apply polygon refinement to the masks. This will smooth the edges of the masks slightly and ensure that each mask consists of a single closed polygon (before merging). - * @default true - */ - apply_polygon_refinement?: boolean; - /** - * Mask Filter - * @description The filtering to apply to the detected masks before merging them into a final output. - * @default all - * @enum {string} - */ - mask_filter?: "all" | "largest" | "highest_box_score"; + strings?: string[] | null; /** * type - * @default segment_anything + * @default string_batch * @constant */ - type: "segment_anything"; - }; - /** SessionProcessorStatus */ - SessionProcessorStatus: { - /** - * Is Started - * @description Whether the session processor is started - */ - is_started: boolean; - /** - * Is Processing - * @description Whether a session is being processed - */ - is_processing: boolean; + type: "string_batch"; }; /** - * SessionQueueAndProcessorStatus - * @description The overall status of session queue and processor + * String Collection Index + * @description CollectionIndex Picks an index out of a collection with a random option */ - SessionQueueAndProcessorStatus: { - queue: components["schemas"]["SessionQueueStatus"]; - processor: components["schemas"]["SessionProcessorStatus"]; - }; - /** SessionQueueCountsByDestination */ - SessionQueueCountsByDestination: { - /** - * Queue Id - * @description The ID of the queue - */ - queue_id: string; - /** - * Destination - * @description The destination of queue items included in this status - */ - destination: string; - /** - * Pending - * @description Number of queue items with status 'pending' for the destination - */ - pending: number; - /** - * In Progress - * @description Number of queue items with status 'in_progress' for the destination - */ - in_progress: number; + StringCollectionIndexInvocation: { /** - * Completed - * @description Number of queue items with status 'complete' for the destination + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - completed: number; + id: string; /** - * Failed - * @description Number of queue items with status 'error' for the destination + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - failed: number; + is_intermediate?: boolean; /** - * Canceled - * @description Number of queue items with status 'canceled' for the destination + * Use Cache + * @description Whether or not to use the cache + * @default false */ - canceled: number; + use_cache?: boolean; /** - * Total - * @description Total number of queue items for the destination + * Random + * @description Random Index? + * @default true */ - total: number; - }; - /** - * SessionQueueItem - * @description Session queue item without the full graph. Used for serialization. - */ - SessionQueueItem: { + random?: boolean; /** - * Item Id - * @description The identifier of the session queue item + * Index + * @description zero based index into collection (note index will wrap around if out of bounds) + * @default 0 */ - item_id: number; + index?: number; /** - * Status - * @description The status of this queue item - * @default pending - * @enum {string} + * Collection + * @description string collection + * @default null */ - status: "pending" | "in_progress" | "completed" | "failed" | "canceled"; + collection?: string[] | null; /** - * Priority - * @description The priority of this queue item - * @default 0 + * type + * @default string_collection_index + * @constant */ - priority: number; + type: "string_collection_index"; + }; + /** + * String Collection Primitive + * @description A collection of string primitive values + */ + StringCollectionInvocation: { /** - * Batch Id - * @description The ID of the batch associated with this queue item + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - batch_id: string; + id: string; /** - * Origin - * @description The origin of this queue item. This data is used by the frontend to determine how to handle results. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - origin?: string | null; + is_intermediate?: boolean; /** - * Destination - * @description The origin of this queue item. This data is used by the frontend to determine how to handle results + * Use Cache + * @description Whether or not to use the cache + * @default true */ - destination?: string | null; + use_cache?: boolean; /** - * Session Id - * @description The ID of the session associated with this queue item. The session doesn't exist in graph_executions until the queue item is executed. + * Collection + * @description The collection of string values + * @default [] */ - session_id: string; + collection?: string[]; /** - * Error Type - * @description The error type if this queue item errored + * type + * @default string_collection + * @constant */ - error_type?: string | null; + type: "string_collection"; + }; + /** + * String Collection Joiner + * @description Takes a collection of strings and returns a single string with all the collections items, separated by the input delimiter. + */ + StringCollectionJoinerInvocation: { /** - * Error Message - * @description The error message if this queue item errored + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - error_message?: string | null; + id: string; /** - * Error Traceback - * @description The error traceback if this queue item errored + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - error_traceback?: string | null; + is_intermediate?: boolean; /** - * Created At - * @description When this queue item was created + * Use Cache + * @description Whether or not to use the cache + * @default true */ - created_at: string; + use_cache?: boolean; /** - * Updated At - * @description When this queue item was updated + * Delimiter + * @description The character to place between each string. + * @default , */ - updated_at: string; + delimiter?: string; /** - * Started At - * @description When this queue item was started + * Collection + * @description The string collection to join. + * @default null */ - started_at?: string | null; + collection?: string[] | null; /** - * Completed At - * @description When this queue item was completed + * Escape Delimiter + * @description Wehter we should escape the delimiter + * @default false */ - completed_at?: string | null; + escape_delimiter?: boolean; /** - * Queue Id - * @description The id of the queue with which this item is associated + * type + * @default string_collection_joiner_invocation + * @constant */ - queue_id: string; + type: "string_collection_joiner_invocation"; + }; + /** + * StringCollectionJoinerOutput + * @description String Collection Joiner Output + */ + StringCollectionJoinerOutput: { /** - * User Id - * @description The id of the user who created this queue item - * @default system + * Result + * @description The joined string */ - user_id?: string; + result: string; /** - * User Display Name - * @description The display name of the user who created this queue item, if available + * type + * @default string_collection_joiner_output + * @constant */ - user_display_name?: string | null; + type: "string_collection_joiner_output"; + }; + /** + * String Collection Primitive Linked + * @description Allows creation of collection and optionally add a collection + */ + StringCollectionLinkedInvocation: { /** - * User Email - * @description The email of the user who created this queue item, if available + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - user_email?: string | null; + id: string; /** - * Field Values - * @description The field values that were used for this queue item + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - field_values?: components["schemas"]["NodeFieldValue"][] | null; + is_intermediate?: boolean; /** - * Retried From Item Id - * @description The item_id of the queue item that this item was retried from + * Use Cache + * @description Whether or not to use the cache + * @default true */ - retried_from_item_id?: number | null; - /** @description The fully-populated session to be executed */ - session: components["schemas"]["GraphExecutionState"]; - /** @description The workflow associated with this queue item */ - workflow?: components["schemas"]["WorkflowWithoutID"] | null; - }; - /** SessionQueueStatus */ - SessionQueueStatus: { + use_cache?: boolean; /** - * Queue Id - * @description The ID of the queue + * Collection + * @description The collection of string values + * @default [] */ - queue_id: string; + collection?: string[]; /** - * Item Id - * @description The current queue item id + * type + * @default string_collection_linked + * @constant */ - item_id: number | null; + type: "string_collection_linked"; /** - * Batch Id - * @description The current queue item's batch id + * Value + * @description The string value + * @default null */ - batch_id: string | null; + value?: string | null; + }; + /** + * StringCollectionOutput + * @description Base class for nodes that output a collection of strings + */ + StringCollectionOutput: { /** - * Session Id - * @description The current queue item's session id + * Collection + * @description The output strings */ - session_id: string | null; + collection: string[]; /** - * Pending - * @description Number of queue items with status 'pending' + * type + * @default string_collection_output + * @constant */ - pending: number; + type: "string_collection_output"; + }; + /** + * String Collection Toggle + * @description Allows boolean selection between two separate string collection inputs + */ + StringCollectionToggleInvocation: { /** - * In Progress - * @description Number of queue items with status 'in_progress' + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - in_progress: number; + id: string; /** - * Completed - * @description Number of queue items with status 'complete' + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - completed: number; + is_intermediate?: boolean; /** - * Failed - * @description Number of queue items with status 'error' + * Use Cache + * @description Whether or not to use the cache + * @default true */ - failed: number; + use_cache?: boolean; /** - * Canceled - * @description Number of queue items with status 'canceled' + * Use Second + * @description Use 2nd Input + * @default false */ - canceled: number; + use_second?: boolean; /** - * Total - * @description Total number of queue items + * Col1 + * @description First String Collection Input + * @default null */ - total: number; + col1?: string[] | null; /** - * User Pending - * @description Number of queue items with status 'pending' for the current user + * Col2 + * @description Second String Collection Input + * @default null */ - user_pending?: number | null; + col2?: string[] | null; /** - * User In Progress - * @description Number of queue items with status 'in_progress' for the current user + * type + * @default string_collection_toggle + * @constant */ - user_in_progress?: number | null; + type: "string_collection_toggle"; }; /** - * SetupRequest - * @description Request body for initial admin setup. + * String Generator + * @description Generated a range of strings for use in a batched generation */ - SetupRequest: { + StringGenerator: { /** - * Email - * @description Admin email address + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - email: string; + id: string; /** - * Display Name - * @description Admin display name + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - display_name?: string | null; + is_intermediate?: boolean; /** - * Password - * @description Admin password + * Use Cache + * @description Whether or not to use the cache + * @default true */ - password: string; - }; - /** - * SetupResponse - * @description Response from successful admin setup. - */ - SetupResponse: { + use_cache?: boolean; /** - * Success - * @description Whether setup was successful + * Generator Type + * @description The string generator. + */ + generator: components["schemas"]["StringGeneratorField"]; + /** + * type + * @default string_generator + * @constant */ - success: boolean; - /** @description Created admin user information */ - user: components["schemas"]["UserDTO"]; + type: "string_generator"; }; + /** StringGeneratorField */ + StringGeneratorField: Record; /** - * SetupStatusResponse - * @description Response for setup status check. + * StringGeneratorOutput + * @description Base class for nodes that output a collection of strings */ - SetupStatusResponse: { - /** - * Setup Required - * @description Whether initial setup is required - */ - setup_required: boolean; + StringGeneratorOutput: { /** - * Multiuser Enabled - * @description Whether multiuser mode is enabled + * Strings + * @description The generated strings */ - multiuser_enabled: boolean; + strings: string[]; /** - * Strict Password Checking - * @description Whether strict password requirements are enforced + * type + * @default string_generator_output + * @constant */ - strict_password_checking: boolean; + type: "string_generator_output"; }; /** - * Show Image - * @description Displays a provided image using the OS image viewer, and passes it forward in the pipeline. + * String Primitive + * @description A string primitive value */ - ShowImageInvocation: { + StringInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -24635,110 +38639,150 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to show - * @default null + * Value + * @description The string value + * @default */ - image?: components["schemas"]["ImageField"] | null; + value?: string; /** * type - * @default show_image + * @default string * @constant */ - type: "show_image"; + type: "string"; }; /** - * SigLIP_Diffusers_Config - * @description Model config for SigLIP. + * String Join + * @description Joins string left to string right */ - SigLIP_Diffusers_Config: { + StringJoinInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * String Left + * @description String Left + * @default */ - file_size: number; + string_left?: string; /** - * Name - * @description Name of the model. + * String Right + * @description String Right + * @default */ - name: string; + string_right?: string; /** - * Description - * @description Model description + * type + * @default string_join + * @constant */ - description: string | null; + type: "string_join"; + }; + /** + * String Join Three + * @description Joins string left to string middle to string right + */ + StringJoinThreeInvocation: { /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Format - * @default diffusers - * @constant + * String Left + * @description String Left + * @default */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + string_left?: string; /** - * Type - * @default siglip - * @constant + * String Middle + * @description String Middle + * @default */ - type: "siglip"; + string_middle?: string; /** - * Base - * @default any + * String Right + * @description String Right + * @default + */ + string_right?: string; + /** + * type + * @default string_join_three * @constant */ - base: "any"; + type: "string_join_three"; + }; + /** + * StringOutput + * @description Base class for nodes that output a single string + */ + StringOutput: { /** - * Cpu Only - * @description Whether this model should run on CPU only + * Value + * @description The output string */ - cpu_only: boolean | null; + value: string; + /** + * type + * @default string_output + * @constant + */ + type: "string_output"; }; /** - * Image-to-Image (Autoscale) - * @description Run any spandrel image-to-image model (https://github.com/chaiNNer-org/spandrel) until the target scale is reached. + * StringPosNegOutput + * @description Base class for invocations that output a positive and negative string */ - SpandrelImageToImageAutoscaleInvocation: { + StringPosNegOutput: { /** - * @description The board to save the image to - * @default null + * Positive String + * @description Positive string */ - board?: components["schemas"]["BoardField"] | null; + positive_string: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Negative String + * @description Negative string */ - metadata?: components["schemas"]["MetadataField"] | null; + negative_string: string; + /** + * type + * @default string_pos_neg_output + * @constant + */ + type: "string_pos_neg_output"; + }; + /** + * String Replace + * @description Replaces the search string with the replace string + */ + StringReplaceInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -24757,56 +38801,41 @@ export type components = { */ use_cache?: boolean; /** - * @description The input image - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Image-to-Image Model - * @description Image-to-Image model - * @default null - */ - image_to_image_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Tile Size - * @description The tile size for tiled image-to-image. Set to 0 to disable tiling. - * @default 512 + * String + * @description String to work on + * @default */ - tile_size?: number; + string?: string; /** - * type - * @default spandrel_image_to_image_autoscale - * @constant + * Search String + * @description String to search for + * @default */ - type: "spandrel_image_to_image_autoscale"; + search_string?: string; /** - * Scale - * @description The final scale of the output image. If the model does not upscale the image, this will be ignored. - * @default 4 + * Replace String + * @description String to replace the search + * @default */ - scale?: number; + replace_string?: string; /** - * Fit To Multiple Of 8 - * @description If true, the output image will be resized to the nearest multiple of 8 in both dimensions. + * Use Regex + * @description Use search string as a regex expression (non regex is case insensitive) * @default false */ - fit_to_multiple_of_8?: boolean; - }; - /** - * Image-to-Image - * @description Run any spandrel image-to-image model (https://github.com/chaiNNer-org/spandrel). - */ - SpandrelImageToImageInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null + use_regex?: boolean; + /** + * type + * @default string_replace + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "string_replace"; + }; + /** + * String Split + * @description Splits string into two strings, based on the first occurance of the delimiter. The delimiter will be removed from the string + */ + StringSplitInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -24825,202 +38854,146 @@ export type components = { */ use_cache?: boolean; /** - * @description The input image - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Image-to-Image Model - * @description Image-to-Image model - * @default null + * String + * @description String to split + * @default */ - image_to_image_model?: components["schemas"]["ModelIdentifierField"] | null; + string?: string; /** - * Tile Size - * @description The tile size for tiled image-to-image. Set to 0 to disable tiling. - * @default 512 + * Delimiter + * @description Delimiter to spilt with. blank will split on the first whitespace + * @default */ - tile_size?: number; + delimiter?: string; /** * type - * @default spandrel_image_to_image + * @default string_split * @constant */ - type: "spandrel_image_to_image"; + type: "string_split"; }; /** - * Spandrel_Checkpoint_Config - * @description Model config for Spandrel Image to Image models. + * String Split Negative + * @description Splits string into two strings, inside [] goes into negative string everthing else goes into positive string. Each [ and ] character is replaced with a space */ - Spandrel_Checkpoint_Config: { + StringSplitNegInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * String + * @description String to split + * @default */ - file_size: number; + string?: string; /** - * Name - * @description Name of the model. + * type + * @default string_split_neg + * @constant */ - name: string; + type: "string_split_neg"; + }; + /** + * String to Collection Splitter + * @description Takes a delimited string and splits it into a collection. + */ + StringToCollectionSplitterInvocation: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Delimiter + * @description The character dividing each string. + * @default , */ - cover_image: string | null; + delimiter?: string; /** - * Base - * @default any - * @constant + * String + * @description The string to split. + * @default null */ - base: "any"; + string?: string | null; /** - * Type - * @default spandrel_image_to_image - * @constant + * Escape Delimiter + * @description Whether we should unescape the delimiter + * @default false */ - type: "spandrel_image_to_image"; + unescape_delimiter?: boolean; /** - * Format - * @default checkpoint + * type + * @default string_to_collection_splitter_invocation * @constant */ - format: "checkpoint"; - }; - /** StarredImagesResult */ - StarredImagesResult: { - /** - * Affected Boards - * @description The ids of boards affected by the delete operation - */ - affected_boards: string[]; - /** - * Starred Images - * @description The names of the images that were starred - */ - starred_images: string[]; + type: "string_to_collection_splitter_invocation"; }; - /** StarterModel */ - StarterModel: { - /** Description */ - description: string; - /** Source */ - source: string; - /** Name */ - name: string; - base: components["schemas"]["BaseModelType"]; - type: components["schemas"]["ModelType"]; - format?: components["schemas"]["ModelFormat"] | null; - /** - * Is Installed - * @default false - */ - is_installed?: boolean; + /** + * String To Float + * @description Converts a string to a float + */ + StringToFloatInvocation: { /** - * Previous Names - * @default [] + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - previous_names?: string[]; - /** Dependencies */ - dependencies?: components["schemas"]["StarterModelWithoutDependencies"][] | null; - }; - /** StarterModelBundle */ - StarterModelBundle: { - /** Name */ - name: string; - /** Models */ - models: components["schemas"]["StarterModel"][]; - }; - /** StarterModelResponse */ - StarterModelResponse: { - /** Starter Models */ - starter_models: components["schemas"]["StarterModel"][]; - /** Starter Bundles */ - starter_bundles: { - [key: string]: components["schemas"]["StarterModelBundle"]; - }; - }; - /** StarterModelWithoutDependencies */ - StarterModelWithoutDependencies: { - /** Description */ - description: string; - /** Source */ - source: string; - /** Name */ - name: string; - base: components["schemas"]["BaseModelType"]; - type: components["schemas"]["ModelType"]; - format?: components["schemas"]["ModelFormat"] | null; + id: string; /** - * Is Installed + * Is Intermediate + * @description Whether or not this is an intermediate invocation. * @default false */ - is_installed?: boolean; - /** - * Previous Names - * @default [] - */ - previous_names?: string[]; - }; - /** - * String2Output - * @description Base class for invocations that output two strings - */ - String2Output: { + is_intermediate?: boolean; /** - * String 1 - * @description string 1 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - string_1: string; + use_cache?: boolean; /** - * String 2 - * @description string 2 + * Float String + * @description string containing a float to convert + * @default null */ - string_2: string; + float_string?: string | null; /** * type - * @default string_2_output + * @default string_to_float * @constant */ - type: "string_2_output"; + type: "string_to_float"; }; /** - * String Batch - * @description Create a batched generation, where the workflow is executed once for each string in the batch. + * String To Int + * @description Converts a string to an integer */ - StringBatchInvocation: { + StringToIntInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25039,30 +39012,23 @@ export type components = { */ use_cache?: boolean; /** - * Batch Group - * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. - * @default None - * @enum {string} - */ - batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; - /** - * Strings - * @description The strings to batch over + * Int String + * @description string containing an integer to convert * @default null */ - strings?: string[] | null; + int_string?: string | null; /** * type - * @default string_batch + * @default string_to_int * @constant */ - type: "string_batch"; + type: "string_to_int"; }; /** - * String Collection Primitive - * @description A collection of string primitive values + * String To LoRA + * @description Loads a lora from a json string, outputting its submodels. */ - StringCollectionInvocation: { + StringToLoraInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25081,40 +39047,45 @@ export type components = { */ use_cache?: boolean; /** - * Collection - * @description The collection of string values - * @default [] + * Model String + * @description string containing a Model to convert + * @default null */ - collection?: string[]; + model_string?: string | null; /** * type - * @default string_collection + * @default string_to_lora * @constant */ - type: "string_collection"; + type: "string_to_lora"; }; /** - * StringCollectionOutput - * @description Base class for nodes that output a collection of strings + * StringToLoraOutput + * @description String to Lora model output */ - StringCollectionOutput: { + StringToLoraOutput: { /** - * Collection - * @description The output strings + * Model + * @description LoRA model to load */ - collection: string[]; + model: components["schemas"]["ModelIdentifierField"]; + /** + * Name + * @description Model Name + */ + name: string; /** * type - * @default string_collection_output + * @default string_to_lora_output * @constant */ - type: "string_collection_output"; + type: "string_to_lora_output"; }; /** - * String Generator - * @description Generated a range of strings for use in a batched generation + * String To Main Model + * @description Loads a main model from a json string, outputting its submodels. */ - StringGenerator: { + StringToMainModelInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25133,41 +39104,45 @@ export type components = { */ use_cache?: boolean; /** - * Generator Type - * @description The string generator. + * Model String + * @description string containing a Model to convert + * @default null */ - generator: components["schemas"]["StringGeneratorField"]; + model_string?: string | null; /** * type - * @default string_generator + * @default string_to_main_model * @constant */ - type: "string_generator"; + type: "string_to_main_model"; }; - /** StringGeneratorField */ - StringGeneratorField: Record; /** - * StringGeneratorOutput - * @description Base class for nodes that output a collection of strings + * StringToMainModelOutput + * @description String to main model output */ - StringGeneratorOutput: { + StringToMainModelOutput: { /** - * Strings - * @description The generated strings + * Model + * @description Main model (UNet, VAE, CLIP) to load */ - strings: string[]; + model: components["schemas"]["ModelIdentifierField"]; + /** + * Name + * @description Model Name + */ + name: string; /** * type - * @default string_generator_output + * @default string_to_main_model_output * @constant */ - type: "string_generator_output"; + type: "string_to_main_model_output"; }; /** - * String Primitive - * @description A string primitive value + * String To Model + * @description Loads a model from a json string, outputting its submodels. */ - StringInvocation: { + StringToModelInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25186,64 +39161,45 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The string value - * @default + * Model String + * @description string containing a Model to convert + * @default null */ - value?: string; + model_string?: string | null; /** * type - * @default string + * @default string_to_model * @constant */ - type: "string"; + type: "string_to_model"; }; /** - * String Join - * @description Joins string left to string right + * StringToModelOutput + * @description String to model output */ - StringJoinInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; + StringToModelOutput: { /** - * String Left - * @description String Left - * @default + * Model + * @description Model identifier */ - string_left?: string; + model: components["schemas"]["ModelIdentifierField"]; /** - * String Right - * @description String Right - * @default + * Name + * @description Model Name */ - string_right?: string; + name: string; /** * type - * @default string_join + * @default string_to_model_output * @constant */ - type: "string_join"; + type: "string_to_model_output"; }; /** - * String Join Three - * @description Joins string left to string middle to string right + * String To SDXL Main Model + * @description Loads a SDXL model from a json string, outputting its submodels. */ - StringJoinThreeInvocation: { + StringToSDXLModelInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25262,74 +39218,45 @@ export type components = { */ use_cache?: boolean; /** - * String Left - * @description String Left - * @default - */ - string_left?: string; - /** - * String Middle - * @description String Middle - * @default - */ - string_middle?: string; - /** - * String Right - * @description String Right - * @default - */ - string_right?: string; - /** - * type - * @default string_join_three - * @constant - */ - type: "string_join_three"; - }; - /** - * StringOutput - * @description Base class for nodes that output a single string - */ - StringOutput: { - /** - * Value - * @description The output string + * Model String + * @description string containing a Model to convert + * @default null */ - value: string; + model_string?: string | null; /** * type - * @default string_output + * @default string_to_sdxl_model * @constant */ - type: "string_output"; + type: "string_to_sdxl_model"; }; /** - * StringPosNegOutput - * @description Base class for invocations that output a positive and negative string + * StringToSDXLModelOutput + * @description String to SDXL main model output */ - StringPosNegOutput: { + StringToSDXLModelOutput: { /** - * Positive String - * @description Positive string + * Model + * @description Main model (UNet, VAE, CLIP) to load */ - positive_string: string; + model: components["schemas"]["ModelIdentifierField"]; /** - * Negative String - * @description Negative string + * Name + * @description Model Name */ - negative_string: string; + name: string; /** * type - * @default string_pos_neg_output + * @default string_to_sdxl_model_output * @constant */ - type: "string_pos_neg_output"; + type: "string_to_sdxl_model_output"; }; /** - * String Replace - * @description Replaces the search string with the replace string + * String To Scheduler + * @description Converts a string to a scheduler */ - StringReplaceInvocation: { + StringToSchedulerInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25348,41 +39275,23 @@ export type components = { */ use_cache?: boolean; /** - * String - * @description String to work on - * @default - */ - string?: string; - /** - * Search String - * @description String to search for - * @default - */ - search_string?: string; - /** - * Replace String - * @description String to replace the search - * @default - */ - replace_string?: string; - /** - * Use Regex - * @description Use search string as a regex expression (non regex is case insensitive) - * @default false + * Scheduler String + * @description string containing a scheduler to convert + * @default null */ - use_regex?: boolean; + scheduler_string?: string | null; /** * type - * @default string_replace + * @default string_to_scheduler * @constant */ - type: "string_replace"; + type: "string_to_scheduler"; }; /** - * String Split - * @description Splits string into two strings, based on the first occurance of the delimiter. The delimiter will be removed from the string + * String Toggle + * @description Allows boolean selection between two separate string inputs */ - StringSplitInvocation: { + StringToggleInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25401,29 +39310,35 @@ export type components = { */ use_cache?: boolean; /** - * String - * @description String to split - * @default + * Use Second + * @description Use 2nd Input + * @default false */ - string?: string; + use_second?: boolean; /** - * Delimiter - * @description Delimiter to spilt with. blank will split on the first whitespace - * @default + * String 1 + * @description First String Input + * @default null */ - delimiter?: string; + str1?: string | null; + /** + * String 2 + * @description Second String Input + * @default null + */ + str2?: string | null; /** * type - * @default string_split + * @default string_toggle * @constant */ - type: "string_split"; + type: "string_toggle"; }; /** - * String Split Negative - * @description Splits string into two strings, inside [] goes into negative string everthing else goes into positive string. Each [ and ] character is replaced with a space + * Strings To CSV + * @description Strings To CSV converts a a list of Strings into a CSV */ - StringSplitNegInvocation: { + StringsToCSVInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25438,21 +39353,21 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * String - * @description String to split + * Strings + * @description String or Collection of Strings to convert to CSV format * @default */ - string?: string; + strings?: string | string[]; /** * type - * @default string_split_neg + * @default strings_to_csv * @constant */ - type: "string_split_neg"; + type: "strings_to_csv"; }; /** * StylePresetField @@ -25540,14 +39455,73 @@ export type components = { * @default sub * @constant */ - type: "sub"; - }; - /** T2IAdapterField */ - T2IAdapterField: { - /** @description The T2I-Adapter image prompt. */ - image: components["schemas"]["ImageField"]; - /** @description The T2I-Adapter model to use. */ - t2i_adapter_model: components["schemas"]["ModelIdentifierField"]; + type: "sub"; + }; + /** T2IAdapterField */ + T2IAdapterField: { + /** @description The T2I-Adapter image prompt. */ + image: components["schemas"]["ImageField"]; + /** @description The T2I-Adapter model to use. */ + t2i_adapter_model: components["schemas"]["ModelIdentifierField"]; + /** + * Weight + * @description The weight given to the T2I-Adapter + * @default 1 + */ + weight?: number | number[]; + /** + * Begin Step Percent + * @description When the T2I-Adapter is first applied (% of total steps) + * @default 0 + */ + begin_step_percent?: number; + /** + * End Step Percent + * @description When the T2I-Adapter is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; + /** + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} + */ + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + }; + /** + * T2I-Adapter - SD1.5, SDXL + * @description Collects T2I-Adapter info to pass to other nodes. + */ + T2IAdapterInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The IP-Adapter image prompt. + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * T2I-Adapter Model + * @description The T2I-Adapter model. + * @default null + */ + t2i_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; /** * Weight * @description The weight given to the T2I-Adapter @@ -25568,17 +39542,23 @@ export type components = { end_step_percent?: number; /** * Resize Mode - * @description The resize mode to use + * @description The resize mode applied to the T2I-Adapter input image so that it matches the target output size. * @default just_resize * @enum {string} */ resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + /** + * type + * @default t2i_adapter + * @constant + */ + type: "t2i_adapter"; }; /** - * T2I-Adapter - SD1.5, SDXL + * T2I-Adapter-Linked * @description Collects T2I-Adapter info to pass to other nodes. */ - T2IAdapterInvocation: { + T2IAdapterLinkedInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -25634,10 +39614,30 @@ export type components = { resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** * type - * @default t2i_adapter + * @default t2i_adapter_linked * @constant */ - type: "t2i_adapter"; + type: "t2i_adapter_linked"; + /** + * T2I Adapter List + * @description T2I-Adapter(s) to apply + * @default null + */ + t2i_adapter_list?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + }; + /** T2IAdapterListOutput */ + T2IAdapterListOutput: { + /** + * T2I Adapter List + * @description T2I-Adapter(s) to apply + */ + t2i_adapter_list: components["schemas"]["T2IAdapterField"][]; + /** + * type + * @default t2i_adapter_list_output + * @constant + */ + type: "t2i_adapter_list_output"; }; /** T2IAdapterMetadataField */ T2IAdapterMetadataField: { @@ -25973,43 +39973,247 @@ export type components = { * @description Url for image to preview model */ cover_image: string | null; - /** - * Base - * @default any - * @constant - */ - base: "any"; + /** + * Base + * @default any + * @constant + */ + base: "any"; + /** + * Type + * @default t5_encoder + * @constant + */ + type: "t5_encoder"; + /** + * Format + * @default t5_encoder + * @constant + */ + format: "t5_encoder"; + /** + * Cpu Only + * @description Whether this model should run on CPU only + */ + cpu_only: boolean | null; + }; + /** TBLR */ + TBLR: { + /** Top */ + top: number; + /** Bottom */ + bottom: number; + /** Left */ + left: number; + /** Right */ + right: number; + }; + /** TI_File_SD1_Config */ + TI_File_SD1_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default embedding + * @constant + */ + type: "embedding"; + /** + * Format + * @default embedding_file + * @constant + */ + format: "embedding_file"; + /** + * Base + * @default sd-1 + * @constant + */ + base: "sd-1"; + }; + /** TI_File_SD2_Config */ + TI_File_SD2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default embedding + * @constant + */ + type: "embedding"; + /** + * Format + * @default embedding_file + * @constant + */ + format: "embedding_file"; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; + }; + /** TI_File_SDXL_Config */ + TI_File_SDXL_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; /** * Type - * @default t5_encoder + * @default embedding * @constant */ - type: "t5_encoder"; + type: "embedding"; /** * Format - * @default t5_encoder + * @default embedding_file * @constant */ - format: "t5_encoder"; + format: "embedding_file"; /** - * Cpu Only - * @description Whether this model should run on CPU only + * Base + * @default sdxl + * @constant */ - cpu_only: boolean | null; - }; - /** TBLR */ - TBLR: { - /** Top */ - top: number; - /** Bottom */ - bottom: number; - /** Left */ - left: number; - /** Right */ - right: number; + base: "sdxl"; }; - /** TI_File_SD1_Config */ - TI_File_SD1_Config: { + /** TI_Folder_SD1_Config */ + TI_Folder_SD1_Config: { /** * Key * @description A unique key for this model. @@ -26065,10 +40269,10 @@ export type components = { type: "embedding"; /** * Format - * @default embedding_file + * @default embedding_folder * @constant */ - format: "embedding_file"; + format: "embedding_folder"; /** * Base * @default sd-1 @@ -26076,8 +40280,8 @@ export type components = { */ base: "sd-1"; }; - /** TI_File_SD2_Config */ - TI_File_SD2_Config: { + /** TI_Folder_SD2_Config */ + TI_Folder_SD2_Config: { /** * Key * @description A unique key for this model. @@ -26133,10 +40337,10 @@ export type components = { type: "embedding"; /** * Format - * @default embedding_file + * @default embedding_folder * @constant */ - format: "embedding_file"; + format: "embedding_folder"; /** * Base * @default sd-2 @@ -26144,8 +40348,8 @@ export type components = { */ base: "sd-2"; }; - /** TI_File_SDXL_Config */ - TI_File_SDXL_Config: { + /** TI_Folder_SDXL_Config */ + TI_Folder_SDXL_Config: { /** * Key * @description A unique key for this model. @@ -26201,10 +40405,10 @@ export type components = { type: "embedding"; /** * Format - * @default embedding_file + * @default embedding_folder * @constant */ - format: "embedding_file"; + format: "embedding_folder"; /** * Base * @default sdxl @@ -26212,220 +40416,469 @@ export type components = { */ base: "sdxl"; }; - /** TI_Folder_SD1_Config */ - TI_Folder_SD1_Config: { + /** + * TensorField + * @description A tensor primitive field. + */ + TensorField: { /** - * Key - * @description A unique key for this model. + * Tensor Name + * @description The name of a tensor. */ - key: string; + tensor_name: string; + }; + /** + * Text Mask + * @description Creates a 2D rendering of a text mask from a given font. + * + * Create a white on black (or black on white) text image for use with controlnets or further processing in other nodes. Specify any TTF/OTF font file available to Invoke and control parameters to resize, rotate, and reposition the text. + * + * Currently this only generates one line of text, but it can be layered with other images using the Image Compositor node or any other such tool. + */ + TextMaskInvocation: { /** - * Hash - * @description The hash of the model file(s). + * @description The board to save the image to + * @default null */ - hash: string; + board?: components["schemas"]["BoardField"] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * @description Optional metadata to be saved with the image + * @default null */ - path: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * File Size - * @description The size of the model in bytes. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - file_size: number; + id: string; /** - * Name - * @description Name of the model. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Width + * @description The width of the desired mask + * @default 512 + */ + width?: number; + /** + * Height + * @description The height of the desired mask + * @default 512 + */ + height?: number; + /** + * Text + * @description The text to render + * @default + */ + text?: string; + /** + * Font + * @description Path to a FreeType-supported TTF/OTF font file + * @default + */ + font?: string; + /** + * Size + * @description Desired point size of text to use + * @default 64 + */ + size?: number; + /** + * Angle + * @description Angle of rotation to apply to the text + * @default 0 + */ + angle?: number; + /** + * X Offset + * @description x-offset for text rendering + * @default 24 + */ + x_offset?: number; + /** + * Y Offset + * @description y-offset for text rendering + * @default 36 + */ + y_offset?: number; + /** + * Invert + * @description Whether to invert color of the output + * @default false + */ + invert?: boolean; + /** + * type + * @default text_mask + * @constant + */ + type: "text_mask"; + }; + /** + * Text to Mask Advanced (Clipseg) + * @description Uses the Clipseg model to generate an image mask from a text prompt. + * + * Output up to four prompt masks combined with logical "and", logical "or", or as separate channels of an RGBA image. + */ + TextToMaskClipsegAdvancedInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image from which to create a mask + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Invert Output + * @description Off: white on black / On: black on white + * @default true + */ + invert_output?: boolean; + /** + * Prompt 1 + * @description First prompt with which to create a mask + * @default null + */ + prompt_1?: string | null; + /** + * Prompt 2 + * @description Second prompt with which to create a mask (optional) + * @default null + */ + prompt_2?: string | null; + /** + * Prompt 3 + * @description Third prompt with which to create a mask (optional) + * @default null + */ + prompt_3?: string | null; + /** + * Prompt 4 + * @description Fourth prompt with which to create a mask (optional) + * @default null + */ + prompt_4?: string | null; + /** + * Combine + * @description How to combine the results + * @default or + * @enum {string} + */ + combine?: "or" | "and" | "butnot" | "none (rgba multiplex)"; + /** + * Smoothing + * @description Radius of blur to apply before thresholding + * @default 4 + */ + smoothing?: number; + /** + * Subject Threshold + * @description Threshold above which is considered the subject + * @default 1 + */ + subject_threshold?: number; + /** + * Background Threshold + * @description Threshold below which is considered the background + * @default 0 + */ + background_threshold?: number; + /** + * type + * @default txt2mask_clipseg_adv + * @constant + */ + type: "txt2mask_clipseg_adv"; + }; + /** + * Text to Mask (Clipseg) + * @description Uses the Clipseg model to generate an image mask from a text prompt. + * + * Input a prompt and an image to generate a mask representing areas of the image matched by the prompt. + */ + TextToMaskClipsegInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The image from which to create a mask + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Invert Output + * @description Off: white on black / On: black on white + * @default true */ - name: string; + invert_output?: boolean; /** - * Description - * @description Model description + * Prompt + * @description The prompt with which to create a mask + * @default null */ - description: string | null; + prompt?: string | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Smoothing + * @description Radius of blur to apply before thresholding + * @default 4 */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + smoothing?: number; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Subject Threshold + * @description Threshold above which is considered the subject + * @default 0.4 */ - source_api_response: string | null; + subject_threshold?: number; /** - * Cover Image - * @description Url for image to preview model + * Background Threshold + * @description Threshold below which is considered the background + * @default 0.4 */ - cover_image: string | null; + background_threshold?: number; /** - * Type - * @default embedding - * @constant + * Mask Expand Or Contract + * @description Pixels by which to grow (or shrink) mask after thresholding + * @default 0 */ - type: "embedding"; + mask_expand_or_contract?: number; /** - * Format - * @default embedding_folder - * @constant + * Mask Blur + * @description Radius of blur to apply after thresholding + * @default 0 */ - format: "embedding_folder"; + mask_blur?: number; /** - * Base - * @default sd-1 + * type + * @default txt2mask_clipseg * @constant */ - base: "sd-1"; + type: "txt2mask_clipseg"; }; - /** TI_Folder_SD2_Config */ - TI_Folder_SD2_Config: { + /** + * Text Font to Image + * @description Turn Text into an image + */ + TextfontimageInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default false */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Text Input + * @description The text from which to generate an image + * @default Invoke AI */ - file_size: number; + text_input?: string; /** - * Name - * @description Name of the model. + * Text Input Second Row + * @description The second row of text to add below the first text + * @default null */ - name: string; + text_input_second_row?: string | null; /** - * Description - * @description Model description + * Second Row Font Size + * @description Font size for the second row of text (optional) + * @default 35 */ - description: string | null; + second_row_font_size?: number | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Font Url + * @description URL address of the font file to download + * @default https://www.1001fonts.com/download/font/caliban.medium.ttf */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + font_url?: string | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Local Font Path + * @description Local font file path (overrides font_url) + * @default null */ - source_api_response: string | null; + local_font_path?: string | null; /** - * Cover Image - * @description Url for image to preview model + * Local Font + * @description Name of the local font file to use from the font_cache folder + * @default None + * @constant */ - cover_image: string | null; + local_font?: "None"; /** - * Type - * @default embedding - * @constant + * Image Width + * @description Width of the output image + * @default 1024 */ - type: "embedding"; + image_width?: number; /** - * Format - * @default embedding_folder - * @constant + * Image Height + * @description Height of the output image + * @default 512 */ - format: "embedding_folder"; + image_height?: number; /** - * Base - * @default sd-2 - * @constant + * Padding + * @description Padding around the text in pixels + * @default 100 */ - base: "sd-2"; - }; - /** TI_Folder_SDXL_Config */ - TI_Folder_SDXL_Config: { + padding?: number; /** - * Key - * @description A unique key for this model. + * Row Gap + * @description Gap between the two rows of text in pixels + * @default 50 */ - key: string; + row_gap?: number; /** - * Hash - * @description The hash of the model file(s). + * type + * @default Text_Font_to_Image + * @constant */ - hash: string; + type: "Text_Font_to_Image"; + }; + /** + * Thresholding + * @description Puts out 3 masks for a source image representing highlights, midtones, and shadows + */ + ThresholdingInvocation: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * @description The board to save the image to + * @default null */ - path: string; + board?: components["schemas"]["BoardField"] | null; /** - * File Size - * @description The size of the model in bytes. + * @description Optional metadata to be saved with the image + * @default null */ - file_size: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Name - * @description Name of the model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - name: string; + id: string; /** - * Description - * @description Model description + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - description: string | null; + is_intermediate?: boolean; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + use_cache?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description The image to add film grain to + * @default null */ - source_api_response: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Cover Image - * @description Url for image to preview model + * Highlights Point + * @description Highlight point + * @default 170 */ - cover_image: string | null; + highlights_point?: number; /** - * Type - * @default embedding - * @constant + * Shadows Point + * @description Shadow point + * @default 85 */ - type: "embedding"; + shadows_point?: number; /** - * Format - * @default embedding_folder - * @constant + * Lut Blur + * @description LUT blur + * @default 0 */ - format: "embedding_folder"; + lut_blur?: number; /** - * Base - * @default sdxl + * type + * @default thresholding * @constant */ - base: "sdxl"; + type: "thresholding"; }; /** - * TensorField - * @description A tensor primitive field. + * ThresholdingOutput + * @description Thresholding output class */ - TensorField: { + ThresholdingOutput: { + highlights_mask: components["schemas"]["ImageField"]; + midtones_mask: components["schemas"]["ImageField"]; + shadows_mask: components["schemas"]["ImageField"]; /** - * Tensor Name - * @description The name of a tensor. + * type + * @default thresholding_output + * @constant */ - tensor_name: string; + type: "thresholding_output"; }; /** Tile */ Tile: { @@ -26434,6 +40887,28 @@ export type components = { /** @description The amount of overlap with adjacent tiles on each side of this tile. */ overlap: components["schemas"]["TBLR"]; }; + /** + * TileSizeOutput + * @description Tile Size Output + */ + TileSizeOutput: { + /** + * Tile Width + * @description Tile Width + */ + tile_width: number; + /** + * Tile Height + * @description Tile Height + */ + tile_height: number; + /** + * type + * @default tile_size_output + * @constant + */ + type: "tile_size_output"; + }; /** * Tile to Properties * @description Split a Tile into its individual properties. @@ -26629,30 +41104,111 @@ export type components = { * @default euler * @enum {string} */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + /** + * UNet + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet?: components["schemas"]["UNetField"] | null; + /** + * CFG Rescale Multiplier + * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR + * @default 0 + */ + cfg_rescale_multiplier?: number; + /** + * Control + * @default null + */ + control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + /** + * type + * @default tiled_multi_diffusion_denoise_latents + * @constant + */ + type: "tiled_multi_diffusion_denoise_latents"; + }; + /** + * TilesOutput + * @description Tiles Output + */ + TilesOutput: { + /** + * Tiles + * @description Tiles Collection + */ + tiles: string[]; + /** + * type + * @default tiles_output + * @constant + */ + type: "tiles_output"; + }; + /** + * Tracery + * @description Takes a collection of json grammars and outputs an expanded string. + */ + TraceryInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Seed + * @description A seed for this run. + * @default 42 + */ + seed?: number; + /** + * Grammars + * @description A collection of grammars. + * @default null + */ + grammars?: string[] | null; /** - * UNet - * @description UNet (scheduler, LoRAs) + * Prompt + * @description The prompt to expand * @default null */ - unet?: components["schemas"]["UNetField"] | null; + prompt?: string | null; /** - * CFG Rescale Multiplier - * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR - * @default 0 + * type + * @default tracery_invocation + * @constant */ - cfg_rescale_multiplier?: number; + type: "tracery_invocation"; + }; + /** + * TraceryOutput + * @description Tracery Output + */ + TraceryOutput: { /** - * Control - * @default null + * Result + * @description The expanded string */ - control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + result: string; /** * type - * @default tiled_multi_diffusion_denoise_latents + * @default tracery_output * @constant */ - type: "tiled_multi_diffusion_denoise_latents"; + type: "tracery_output"; }; /** TransformerField */ TransformerField: { @@ -27585,95 +42141,339 @@ export type components = { * @default vae * @constant */ - type: "vae"; + type: "vae"; + /** + * Base + * @default sd-1 + * @constant + */ + base: "sd-1"; + }; + /** VAE_Diffusers_SDXL_Config */ + VAE_Diffusers_SDXL_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + /** + * Type + * @default vae + * @constant + */ + type: "vae"; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; + }; + /** ValidationError */ + ValidationError: { + /** Location */ + loc: (string | number)[]; + /** Message */ + msg: string; + /** Error Type */ + type: string; + }; + /** WeightedStringOutput */ + WeightedStringOutput: { + /** + * Cleaned Text + * @description The input string with weights and parentheses removed + */ + cleaned_text: string; + /** + * Phrases + * @description List of weighted phrases or words + */ + phrases: string[]; + /** + * Weights + * @description Associated weights for each phrase + */ + weights: number[]; + /** + * Positions + * @description Start positions of each phrase in the cleaned string + */ + positions: number[]; + /** + * type + * @default weighted_string_output + * @constant + */ + type: "weighted_string_output"; + }; + /** Workflow */ + Workflow: { + /** + * Name + * @description The name of the workflow. + */ + name: string; + /** + * Author + * @description The author of the workflow. + */ + author: string; + /** + * Description + * @description The description of the workflow. + */ + description: string; + /** + * Version + * @description The version of the workflow. + */ + version: string; + /** + * Contact + * @description The contact of the workflow. + */ + contact: string; + /** + * Tags + * @description The tags of the workflow. + */ + tags: string; + /** + * Notes + * @description The notes of the workflow. + */ + notes: string; + /** + * Exposedfields + * @description The exposed fields of the workflow. + */ + exposedFields: components["schemas"]["ExposedField"][]; + /** @description The meta of the workflow. */ + meta: components["schemas"]["WorkflowMeta"]; + /** + * Nodes + * @description The nodes of the workflow. + */ + nodes: { + [key: string]: components["schemas"]["JsonValue"]; + }[]; + /** + * Edges + * @description The edges of the workflow. + */ + edges: { + [key: string]: components["schemas"]["JsonValue"]; + }[]; + /** + * Form + * @description The form of the workflow. + */ + form?: { + [key: string]: components["schemas"]["JsonValue"]; + } | null; + /** + * Id + * @description The id of the workflow. + */ + id: string; + }; + /** WorkflowAndGraphResponse */ + WorkflowAndGraphResponse: { + /** + * Workflow + * @description The workflow used to generate the image, as stringified JSON + */ + workflow: string | null; + /** + * Graph + * @description The graph used to generate the image, as stringified JSON + */ + graph: string | null; + }; + /** + * WorkflowCategory + * @enum {string} + */ + WorkflowCategory: "user" | "default"; + /** WorkflowMeta */ + WorkflowMeta: { + /** + * Version + * @description The version of the workflow schema. + */ + version: string; + /** @description The category of the workflow (user or default). */ + category: components["schemas"]["WorkflowCategory"]; + }; + /** WorkflowRecordDTO */ + WorkflowRecordDTO: { + /** + * Workflow Id + * @description The id of the workflow. + */ + workflow_id: string; + /** + * Name + * @description The name of the workflow. + */ + name: string; + /** + * Created At + * @description The created timestamp of the workflow. + */ + created_at: string; + /** + * Updated At + * @description The updated timestamp of the workflow. + */ + updated_at: string; + /** + * Opened At + * @description The opened timestamp of the workflow. + */ + opened_at?: string | null; + /** @description The workflow. */ + workflow: components["schemas"]["Workflow"]; + }; + /** WorkflowRecordListItemWithThumbnailDTO */ + WorkflowRecordListItemWithThumbnailDTO: { + /** + * Workflow Id + * @description The id of the workflow. + */ + workflow_id: string; /** - * Base - * @default sd-1 - * @constant + * Name + * @description The name of the workflow. */ - base: "sd-1"; - }; - /** VAE_Diffusers_SDXL_Config */ - VAE_Diffusers_SDXL_Config: { + name: string; /** - * Key - * @description A unique key for this model. + * Created At + * @description The created timestamp of the workflow. */ - key: string; + created_at: string; /** - * Hash - * @description The hash of the model file(s). + * Updated At + * @description The updated timestamp of the workflow. */ - hash: string; + updated_at: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Opened At + * @description The opened timestamp of the workflow. */ - path: string; + opened_at?: string | null; /** - * File Size - * @description The size of the model in bytes. + * Description + * @description The description of the workflow. */ - file_size: number; + description: string; + /** @description The description of the workflow. */ + category: components["schemas"]["WorkflowCategory"]; /** - * Name - * @description Name of the model. + * Tags + * @description The tags of the workflow. */ - name: string; + tags: string; /** - * Description - * @description Model description + * Thumbnail Url + * @description The URL of the workflow thumbnail. */ - description: string | null; + thumbnail_url?: string | null; + }; + /** + * WorkflowRecordOrderBy + * @description The order by options for workflow records + * @enum {string} + */ + WorkflowRecordOrderBy: "created_at" | "updated_at" | "opened_at" | "name"; + /** WorkflowRecordWithThumbnailDTO */ + WorkflowRecordWithThumbnailDTO: { /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Workflow Id + * @description The id of the workflow. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + workflow_id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Name + * @description The name of the workflow. */ - source_api_response: string | null; + name: string; /** - * Cover Image - * @description Url for image to preview model + * Created At + * @description The created timestamp of the workflow. */ - cover_image: string | null; + created_at: string; /** - * Format - * @default diffusers - * @constant + * Updated At + * @description The updated timestamp of the workflow. */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + updated_at: string; /** - * Type - * @default vae - * @constant + * Opened At + * @description The opened timestamp of the workflow. */ - type: "vae"; + opened_at?: string | null; + /** @description The workflow. */ + workflow: components["schemas"]["Workflow"]; /** - * Base - * @default sdxl - * @constant + * Thumbnail Url + * @description The URL of the workflow thumbnail. */ - base: "sdxl"; - }; - /** ValidationError */ - ValidationError: { - /** Location */ - loc: (string | number)[]; - /** Message */ - msg: string; - /** Error Type */ - type: string; + thumbnail_url?: string | null; }; - /** Workflow */ - Workflow: { + /** WorkflowWithoutID */ + WorkflowWithoutID: { /** * Name * @description The name of the workflow. @@ -27724,234 +42524,424 @@ export type components = { [key: string]: components["schemas"]["JsonValue"]; }[]; /** - * Edges - * @description The edges of the workflow. + * Edges + * @description The edges of the workflow. + */ + edges: { + [key: string]: components["schemas"]["JsonValue"]; + }[]; + /** + * Form + * @description The form of the workflow. + */ + form?: { + [key: string]: components["schemas"]["JsonValue"]; + } | null; + }; + /** + * XY Expand + * @description Takes an XY Item and outputs the X and Y as individual strings + */ + XYExpandInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Xy Item + * @description The XY Item + * @default null + */ + xy_item?: string | null; + /** + * type + * @default xy_expand + * @constant + */ + type: "xy_expand"; + }; + /** + * XYExpandOutput + * @description Two strings that are expanded from an XY Item + */ + XYExpandOutput: { + /** + * X Item + * @description The X item + */ + x_item: string; + /** + * Y Item + * @description The y item + */ + y_item: string; + /** + * type + * @default xy_expand_output + * @constant + */ + type: "xy_expand_output"; + }; + /** + * XYImage Collect + * @description Takes xItem, yItem and an Image and outputs it as an XYImage Item (x_item,y_item,image_name)array converted to json + */ + XYImageCollectInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * X Item + * @description The X item + * @default null + */ + x_item?: string | null; + /** + * Y Item + * @description The Y item + * @default null + */ + y_item?: string | null; + /** + * @description The image to turn into grids + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * type + * @default xy_image_collect + * @constant + */ + type: "xy_image_collect"; + }; + /** + * XYImage Expand + * @description Takes an XYImage item and outputs the XItem,YItem, Image, width & height + */ + XYImageExpandInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Xyimage Item + * @description The XYImage collection item + * @default null + */ + xyimage_item?: string | null; + /** + * type + * @default xy_image_expand + * @constant + */ + type: "xy_image_expand"; + }; + /** + * XYImageExpandOutput + * @description XY Image Expand Output + */ + XYImageExpandOutput: { + /** + * X Item + * @description The X item + */ + x_item: string; + /** + * Y Item + * @description The y item + */ + y_item: string; + /** @description The Image item */ + image: components["schemas"]["ImageField"]; + /** + * Width + * @description The width of the image in pixels + */ + width: number; + /** + * Height + * @description The height of the image in pixels + */ + height: number; + /** + * type + * @default xy_image_expand_output + * @constant + */ + type: "xy_image_expand_output"; + }; + /** + * XYImage Tiles To Image + * @description Takes a collection of XYImage Tiles (json of array(x_pos,y_pos,image_name)) and create an image from overlapping tiles + */ + XYImageTilesToImageInvocation: { + /** + * @description The board to save the image to + * @default null */ - edges: { - [key: string]: components["schemas"]["JsonValue"]; - }[]; + board?: components["schemas"]["BoardField"] | null; /** - * Form - * @description The form of the workflow. + * @description Optional metadata to be saved with the image + * @default null */ - form?: { - [key: string]: components["schemas"]["JsonValue"]; - } | null; + metadata?: components["schemas"]["MetadataField"] | null; /** * Id - * @description The id of the workflow. + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ id: string; - }; - /** WorkflowAndGraphResponse */ - WorkflowAndGraphResponse: { - /** - * Workflow - * @description The workflow used to generate the image, as stringified JSON - */ - workflow: string | null; /** - * Graph - * @description The graph used to generate the image, as stringified JSON + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - graph: string | null; - }; - /** - * WorkflowCategory - * @enum {string} - */ - WorkflowCategory: "user" | "default"; - /** WorkflowMeta */ - WorkflowMeta: { + is_intermediate?: boolean; /** - * Version - * @description The version of the workflow schema. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - version: string; - /** @description The category of the workflow (user or default). */ - category: components["schemas"]["WorkflowCategory"]; - }; - /** WorkflowRecordDTO */ - WorkflowRecordDTO: { + use_cache?: boolean; /** - * Workflow Id - * @description The id of the workflow. + * Xyimages + * @description The xyImage Collection + * @default [] */ - workflow_id: string; + xyimages?: string[]; /** - * Name - * @description The name of the workflow. + * Blend Mode + * @description Seam blending type Linear or Smart + * @default seam-grad + * @enum {string} */ - name: string; + blend_mode?: "Linear" | "seam-grad" | "seam-sobel1" | "seam-sobel3" | "seam-sobel5" | "seam-sobel7" | "seam-scharr"; /** - * Created At - * @description The created timestamp of the workflow. + * Blur Size + * @description Size of the blur & Gutter to use with Smart Seam + * @default 16 */ - created_at: string; + blur_size?: number; /** - * Updated At - * @description The updated timestamp of the workflow. + * Search Size + * @description Seam search size in pixels 1-4 are sensible sizes + * @default 1 */ - updated_at: string; + search_size?: number; /** - * Opened At - * @description The opened timestamp of the workflow. + * type + * @default xy_image_tiles_to_image + * @constant */ - opened_at?: string | null; - /** @description The workflow. */ - workflow: components["schemas"]["Workflow"]; + type: "xy_image_tiles_to_image"; }; - /** WorkflowRecordListItemWithThumbnailDTO */ - WorkflowRecordListItemWithThumbnailDTO: { + /** + * XYImages To Grid + * @description Takes Collection of XYImages (json of (x_item,y_item,image_name)array), sorts the images into X,Y and creates a grid image with labels + */ + XYImagesToGridInvocation: { /** - * Workflow Id - * @description The id of the workflow. + * @description The board to save the image to + * @default null */ - workflow_id: string; + board?: components["schemas"]["BoardField"] | null; /** - * Name - * @description The name of the workflow. + * @description Optional metadata to be saved with the image + * @default null */ - name: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Created At - * @description The created timestamp of the workflow. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - created_at: string; + id: string; /** - * Updated At - * @description The updated timestamp of the workflow. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - updated_at: string; + is_intermediate?: boolean; /** - * Opened At - * @description The opened timestamp of the workflow. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - opened_at?: string | null; + use_cache?: boolean; /** - * Description - * @description The description of the workflow. + * Xyimages + * @description The XYImage item Collection + * @default [] */ - description: string; - /** @description The description of the workflow. */ - category: components["schemas"]["WorkflowCategory"]; + xyimages?: string[]; /** - * Tags - * @description The tags of the workflow. + * Scale Factor + * @description The factor by which to scale the images + * @default 1 */ - tags: string; + scale_factor?: number; /** - * Thumbnail Url - * @description The URL of the workflow thumbnail. + * Resample Mode + * @description The resampling mode + * @default bicubic + * @enum {string} */ - thumbnail_url?: string | null; - }; - /** - * WorkflowRecordOrderBy - * @description The order by options for workflow records - * @enum {string} - */ - WorkflowRecordOrderBy: "created_at" | "updated_at" | "opened_at" | "name"; - /** WorkflowRecordWithThumbnailDTO */ - WorkflowRecordWithThumbnailDTO: { + resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; /** - * Workflow Id - * @description The id of the workflow. + * Left Label Width + * @description Width of the left label area + * @default 100 */ - workflow_id: string; + left_label_width?: number; /** - * Name - * @description The name of the workflow. + * Label Font Size + * @description Size of the font to use for labels + * @default 16 */ - name: string; + label_font_size?: number; /** - * Created At - * @description The created timestamp of the workflow. + * type + * @default xy_images_to_grid + * @constant */ - created_at: string; + type: "xy_images_to_grid"; + }; + /** + * XY Product CSV + * @description Converts X and Y CSV strings to an XY Item collection with every combination of X and Y + */ + XYProductCSVInvocation: { /** - * Updated At - * @description The updated timestamp of the workflow. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - updated_at: string; + id: string; /** - * Opened At - * @description The opened timestamp of the workflow. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - opened_at?: string | null; - /** @description The workflow. */ - workflow: components["schemas"]["Workflow"]; + is_intermediate?: boolean; /** - * Thumbnail Url - * @description The URL of the workflow thumbnail. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - thumbnail_url?: string | null; - }; - /** WorkflowWithoutID */ - WorkflowWithoutID: { + use_cache?: boolean; /** - * Name - * @description The name of the workflow. + * X + * @description x string + * @default null */ - name: string; + x?: string | null; /** - * Author - * @description The author of the workflow. + * Y + * @description y string + * @default null */ - author: string; + y?: string | null; /** - * Description - * @description The description of the workflow. + * type + * @default xy_product_csv + * @constant */ - description: string; + type: "xy_product_csv"; + }; + /** + * XY Product + * @description Takes X and Y string collections and outputs a XY Item collection with every combination of X and Y + */ + XYProductInvocation: { /** - * Version - * @description The version of the workflow. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - version: string; + id: string; /** - * Contact - * @description The contact of the workflow. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - contact: string; + is_intermediate?: boolean; /** - * Tags - * @description The tags of the workflow. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - tags: string; + use_cache?: boolean; /** - * Notes - * @description The notes of the workflow. + * X Collection + * @description The X collection + * @default [] */ - notes: string; + x_collection?: string[]; /** - * Exposedfields - * @description The exposed fields of the workflow. + * Y Collection + * @description The Y collection + * @default [] */ - exposedFields: components["schemas"]["ExposedField"][]; - /** @description The meta of the workflow. */ - meta: components["schemas"]["WorkflowMeta"]; + y_collection?: string[]; /** - * Nodes - * @description The nodes of the workflow. + * type + * @default xy_product + * @constant */ - nodes: { - [key: string]: components["schemas"]["JsonValue"]; - }[]; + type: "xy_product"; + }; + /** + * XYProductOutput + * @description XYCProductOutput a collection that contains every combination of the input collections + */ + XYProductOutput: { /** - * Edges - * @description The edges of the workflow. + * Xy Item Collection + * @description The XY Item collection */ - edges: { - [key: string]: components["schemas"]["JsonValue"]; - }[]; + xy_item_collection: string[]; /** - * Form - * @description The form of the workflow. + * type + * @default xy_collect_output + * @constant */ - form?: { - [key: string]: components["schemas"]["JsonValue"]; - } | null; + type: "xy_collect_output"; }; /** * ZImageConditioningField @@ -29147,6 +44137,10 @@ export interface operations { model_name?: string | null; /** @description Exact match on the format of the model (e.g. 'diffusers') */ model_format?: components["schemas"]["ModelFormat"] | null; + /** @description The field to order by */ + order_by?: components["schemas"]["ModelRecordOrderBy"]; + /** @description The direction to order by */ + direction?: components["schemas"]["SQLiteDirection"]; }; header?: never; path?: never; From 8f89adf71912322ec8a2cfd055c446f97e474dad Mon Sep 17 00:00:00 2001 From: skunkworxdark Date: Sun, 5 Apr 2026 21:24:15 +0100 Subject: [PATCH 5/8] typegen fix - this time without my custom nodes. --- .../frontend/web/src/services/api/schema.ts | 42372 +++++----------- 1 file changed, 13694 insertions(+), 28678 deletions(-) diff --git a/invokeai/frontend/web/src/services/api/schema.ts b/invokeai/frontend/web/src/services/api/schema.ts index d2ad4c8f10d..93c0159ed07 100644 --- a/invokeai/frontend/web/src/services/api/schema.ts +++ b/invokeai/frontend/web/src/services/api/schema.ts @@ -2594,216 +2594,6 @@ export type components = { */ is_active?: boolean | null; }; - /** - * Adv AutoStereogram - * @description create an advanced autostereogram from a depth map - */ - AdvAutostereogramInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The depth map to create the image from - * @default null - */ - depth_map?: components["schemas"]["ImageField"] | null; - /** - * @description The pattern image to use as the background, if not provided then random dots will be used - * @default null - */ - pattern?: components["schemas"]["ImageField"] | null; - /** - * Pattern Width - * @description The pattern width pixels - * @default 100 - */ - pattern_width?: number; - /** - * Depth Steps - * @description The number of depth steps, 30-127 is a good range but should be less than the pattern width - * @default 50 - */ - depth_steps?: number; - /** - * Invert Depth Map - * @description Invert the depth map (difference between crossing and uncrossing eyes) - * @default false - */ - invert_depth_map?: boolean; - /** - * Grayscale - * @description Color or Grayscale output - * @default false - */ - grayscale?: boolean; - /** - * type - * @default adv_autostereogram - * @constant - */ - type: "adv_autostereogram"; - }; - /** - * Advanced Text Font to Image - * @description Overlay Text onto an image or blank canvas. - */ - AdvancedTextFontImageInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Text Input - * @description The text from which to generate an image - * @default Invoke AI - */ - text_input?: string; - /** - * Text Input Second Row - * @description The second row of text to add below the first text - * @default null - */ - text_input_second_row?: string | null; - /** - * Font Url - * @description URL address of the font file to download - * @default https://www.1001fonts.com/download/font/caliban.medium.ttf - */ - font_url?: string | null; - /** - * Local Font Path - * @description Local font file path (overrides font_url) - * @default null - */ - local_font_path?: string | null; - /** - * Local Font - * @description Name of the local font file to use from the font_cache folder - * @default None - * @constant - */ - local_font?: "None"; - /** - * Image Width - * @description Width of the output image - * @default 1024 - */ - image_width?: number; - /** - * Image Height - * @description Height of the output image - * @default 512 - */ - image_height?: number; - /** - * Font Color First - * @description Font color for the first row of text in HEX format (e.g., '#FFFFFF') - * @default #FFFFFF - */ - font_color_first?: string; - /** - * X Position First - * @description X position of the first row of text - * @default 0 - */ - x_position_first?: number; - /** - * Y Position First - * @description Y position of the first row of text - * @default 0 - */ - y_position_first?: number; - /** - * Rotation First - * @description Rotation angle of the first row of text (in degrees) - * @default 0 - */ - rotation_first?: number; - /** - * Font Size First - * @description Font size for the first row of text - * @default 35 - */ - font_size_first?: number | null; - /** - * Font Color Second - * @description Font color for the second row of text in HEX format (e.g., '#FFFFFF') - * @default #FFFFFF - */ - font_color_second?: string; - /** - * X Position Second - * @description X position of the second row of text - * @default 0 - */ - x_position_second?: number; - /** - * Y Position Second - * @description Y position of the second row of text - * @default 0 - */ - y_position_second?: number; - /** - * Rotation Second - * @description Rotation angle of the second row of text (in degrees) - * @default 0 - */ - rotation_second?: number; - /** - * Font Size Second - * @description Font size for the second row of text - * @default 35 - */ - font_size_second?: number | null; - /** - * @description An image to place the text onto - * @default null - */ - input_image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default Advanced_Text_Font_to_Image - * @constant - */ - type: "Advanced_Text_Font_to_Image"; - }; /** * Alpha Mask to Tensor * @description Convert a mask image to a tensor. Opaque regions are 1 and transparent regions are 0. @@ -2844,102 +2634,6 @@ export type components = { */ type: "alpha_mask_to_tensor"; }; - /** - * Anamorphic Streaks - * @description Adds anamorphic streaks to the input image - */ - AnamorphicStreaksInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to streak - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The color to use for streaks - * @default { - * "r": 0, - * "g": 100, - * "b": 255, - * "a": 255 - * } - */ - streak_color?: components["schemas"]["ColorField"]; - /** - * Gamma - * @description Gamma to use for finding highlights - * @default 30 - */ - gamma?: number; - /** - * Blur Radius - * @description Radius to blur highlights - * @default 5 - */ - blur_radius?: number; - /** - * Erosion Kernel Size - * @description Amount to erode blurred highlights - * @default 5 - */ - erosion_kernel_size?: number; - /** - * Erosion Iterations - * @description Number of erosion iterations - * @default 2 - */ - erosion_iterations?: number; - /** - * Streak Intensity - * @description Streak Intensity - * @default 10 - */ - streak_intensity?: number; - /** - * Internal Reflection Strength - * @description Internal reflection strength - * @default 0.3 - */ - internal_reflection_strength?: number; - /** - * Streak Width - * @description Streak width (recommended image width) - * @default 512 - */ - streak_width?: number; - /** - * type - * @default anamorphic_streaks - * @constant - */ - type: "anamorphic_streaks"; - }; AnyModelConfig: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; /** * AppVersion @@ -3066,177 +2760,59 @@ export type components = { type: "apply_mask_to_image"; }; /** - * AutoStereogram - * @description create an autostereogram from a depth map + * BaseMetadata + * @description Adds typing data for discriminated union. */ - AutostereogramInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + BaseMetadata: { /** - * @description Optional metadata to be saved with the image - * @default null + * Name + * @description model's name */ - metadata?: components["schemas"]["MetadataField"] | null; + name: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description discriminator enum property added by openapi-typescript + * @enum {string} */ - id: string; + type: "basemetadata"; + }; + /** + * BaseModelType + * @description An enumeration of base model architectures. For example, Stable Diffusion 1.x, Stable Diffusion 2.x, FLUX, etc. + * + * Every model config must have a base architecture type. + * + * Not all models are associated with a base architecture. For example, CLIP models are their own thing, not related + * to any particular model architecture. To simplify internal APIs and make it easier to work with models, we use a + * fallback/null value `BaseModelType.Any` for these models, instead of making the model base optional. + * @enum {string} + */ + BaseModelType: "any" | "sd-1" | "sd-2" | "sd-3" | "sdxl" | "sdxl-refiner" | "flux" | "flux2" | "cogview4" | "z-image" | "unknown"; + /** Batch */ + Batch: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Batch Id + * @description The ID of the batch */ - is_intermediate?: boolean; + batch_id?: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Origin + * @description The origin of this queue item. This data is used by the frontend to determine how to handle results. */ - use_cache?: boolean; + origin?: string | null; /** - * @description The depth map to create the autostereogram from - * @default null + * Destination + * @description The origin of this queue item. This data is used by the frontend to determine how to handle results */ - depth_map?: components["schemas"]["ImageField"] | null; + destination?: string | null; /** - * @description The pattern image, if not provided then random dots will be used - * @default null + * Data + * @description The batch data collection. */ - pattern?: components["schemas"]["ImageField"] | null; - /** - * Pattern Divisions - * @description How many pattern repeats in output 5-10 is in general a good range. lower = more depth but harder to see - * @default 8 - */ - pattern_divisions?: number; - /** - * Invert Depth Map - * @description Invert the depth map (difference between crossing and uncrossing eyes) - * @default false - */ - invert_depth_map?: boolean; - /** - * Grayscale - * @description Color or Grayscale output - * @default false - */ - grayscale?: boolean; - /** - * type - * @default autostereogram - * @constant - */ - type: "autostereogram"; - }; - /** - * Average Images - * @description Average images - */ - AverageImagesInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Images - * @description The collection of images to average - * @default null - */ - images?: components["schemas"]["ImageField"][] | null; - /** - * Gamma - * @description Gamma for color correcting before/after blending - * @default 2.2 - */ - gamma?: number; - /** - * type - * @default average_images - * @constant - */ - type: "average_images"; - }; - /** - * BaseMetadata - * @description Adds typing data for discriminated union. - */ - BaseMetadata: { - /** - * Name - * @description model's name - */ - name: string; - /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} - */ - type: "basemetadata"; - }; - /** - * BaseModelType - * @description An enumeration of base model architectures. For example, Stable Diffusion 1.x, Stable Diffusion 2.x, FLUX, etc. - * - * Every model config must have a base architecture type. - * - * Not all models are associated with a base architecture. For example, CLIP models are their own thing, not related - * to any particular model architecture. To simplify internal APIs and make it easier to work with models, we use a - * fallback/null value `BaseModelType.Any` for these models, instead of making the model base optional. - * @enum {string} - */ - BaseModelType: "any" | "sd-1" | "sd-2" | "sd-3" | "sdxl" | "sdxl-refiner" | "flux" | "flux2" | "cogview4" | "z-image" | "unknown"; - /** Batch */ - Batch: { - /** - * Batch Id - * @description The ID of the batch - */ - batch_id?: string; - /** - * Origin - * @description The origin of this queue item. This data is used by the frontend to determine how to handle results. - */ - origin?: string | null; - /** - * Destination - * @description The origin of this queue item. This data is used by the frontend to determine how to handle results - */ - destination?: string | null; - /** - * Data - * @description The batch data collection. - */ - data?: components["schemas"]["BatchDatum"][][] | null; - /** @description The graph to initialize the session with */ - graph: components["schemas"]["Graph"]; - /** @description The workflow to initialize the session with */ - workflow?: components["schemas"]["WorkflowWithoutID"] | null; + data?: components["schemas"]["BatchDatum"][][] | null; + /** @description The graph to initialize the session with */ + graph: components["schemas"]["Graph"]; + /** @description The workflow to initialize the session with */ + workflow?: components["schemas"]["WorkflowWithoutID"] | null; /** * Runs * @description Int stating how many times to iterate through all possible batch indices @@ -3842,147 +3418,6 @@ export type components = { */ metadata?: string | null; }; - /** - * Bool Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option - */ - BoolCollectionIndexInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Random - * @description Random Index? - * @default true - */ - random?: boolean; - /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) - * @default 0 - */ - index?: number; - /** - * Collection - * @description bool collection - * @default null - */ - collection?: boolean[] | null; - /** - * type - * @default bool_collection_index - * @constant - */ - type: "bool_collection_index"; - }; - /** - * Bool Collection Toggle - * @description Allows boolean selection between two separate boolean collection inputs - */ - BoolCollectionToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * Col1 - * @description First Bool Collection Input - * @default null - */ - col1?: boolean[] | null; - /** - * Col2 - * @description Second Bool Collection Input - * @default null - */ - col2?: boolean[] | null; - /** - * type - * @default bool_collection_toggle - * @constant - */ - type: "bool_collection_toggle"; - }; - /** - * Bool Toggle - * @description Allows boolean selection between two separate boolean inputs - */ - BoolToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * Bool1 - * @description First Bool Input - * @default false - */ - bool1?: boolean; - /** - * Bool2 - * @description Second Bool Input - * @default false - */ - bool2?: boolean; - /** - * type - * @default bool_toggle - * @constant - */ - type: "bool_toggle"; - }; /** * Boolean Collection Primitive * @description A collection of boolean primitive values @@ -4018,47 +3453,6 @@ export type components = { */ type: "boolean_collection"; }; - /** - * Boolean Collection Primitive Linked - * @description A collection of boolean primitive values - */ - BooleanCollectionLinkedInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The collection of boolean values - * @default [] - */ - collection?: boolean[]; - /** - * type - * @default boolean_collection_linked - * @constant - */ - type: "boolean_collection_linked"; - /** - * Value - * @description The boolean value - * @default false - */ - value?: boolean; - }; /** * BooleanCollectionOutput * @description Base class for nodes that output a collection of booleans @@ -4719,10 +4113,10 @@ export type components = { cpu_only: boolean | null; }; /** - * CMYK Color Separation - * @description Get color images from a base color and two others that subtractively mix to obtain it + * CV2 Infill + * @description Infills transparent areas of an image using OpenCV Inpainting */ - CMYKColorSeparationInvocation: { + CV2InfillInvocation: { /** * @description The board to save the image to * @default null @@ -4751,94 +4145,59 @@ export type components = { */ use_cache?: boolean; /** - * Width - * @description Desired image width - * @default 512 + * @description The image to process + * @default null */ - width?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Height - * @description Desired image height - * @default 512 + * type + * @default infill_cv2 + * @constant */ - height?: number; + type: "infill_cv2"; + }; + /** CacheStats */ + CacheStats: { /** - * C Value - * @description Desired final cyan value + * Hits * @default 0 */ - c_value?: number; - /** - * M Value - * @description Desired final magenta value - * @default 25 - */ - m_value?: number; - /** - * Y Value - * @description Desired final yellow value - * @default 28 - */ - y_value?: number; - /** - * K Value - * @description Desired final black value - * @default 76 - */ - k_value?: number; - /** - * C Split - * @description Desired cyan split point % [0..1.0] - * @default 0.5 - */ - c_split?: number; + hits?: number; /** - * M Split - * @description Desired magenta split point % [0..1.0] - * @default 1 + * Misses + * @default 0 */ - m_split?: number; + misses?: number; /** - * Y Split - * @description Desired yellow split point % [0..1.0] + * High Watermark * @default 0 */ - y_split?: number; + high_watermark?: number; /** - * K Split - * @description Desired black split point % [0..1.0] - * @default 0.5 + * In Cache + * @default 0 */ - k_split?: number; + in_cache?: number; /** - * Profile - * @description CMYK Color Profile - * @default Default - * @enum {string} + * Cleared + * @default 0 */ - profile?: "Default" | "PIL"; + cleared?: number; /** - * type - * @default cmyk_separation - * @constant + * Cache Size + * @default 0 */ - type: "cmyk_separation"; + cache_size?: number; + /** Loaded Model Sizes */ + loaded_model_sizes?: { + [key: string]: number; + }; }; /** - * CMYK Halftone - * @description Halftones an image in the style of a CMYK print + * Calculate Image Tiles Even Split + * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. */ - CMYKHalftoneInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + CalculateImageTilesEvenSplitInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4857,92 +4216,106 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to halftone - * @default null + * Image Width + * @description The image width, in pixels, to calculate tiles for. + * @default 1024 */ - image?: components["schemas"]["ImageField"] | null; + image_width?: number; /** - * Spacing - * @description Halftone dot spacing - * @default 8 + * Image Height + * @description The image height, in pixels, to calculate tiles for. + * @default 1024 */ - spacing?: number; + image_height?: number; /** - * C Angle - * @description C halftone angle - * @default 15 + * Num Tiles X + * @description Number of tiles to divide image into on the x axis + * @default 2 */ - c_angle?: number; + num_tiles_x?: number; /** - * M Angle - * @description M halftone angle - * @default 75 + * Num Tiles Y + * @description Number of tiles to divide image into on the y axis + * @default 2 */ - m_angle?: number; + num_tiles_y?: number; /** - * Y Angle - * @description Y halftone angle - * @default 90 + * Overlap + * @description The overlap, in pixels, between adjacent tiles. + * @default 128 */ - y_angle?: number; + overlap?: number; /** - * K Angle - * @description K halftone angle - * @default 45 + * type + * @default calculate_image_tiles_even_split + * @constant */ - k_angle?: number; + type: "calculate_image_tiles_even_split"; + }; + /** + * Calculate Image Tiles + * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. + */ + CalculateImageTilesInvocation: { /** - * Oversampling - * @description Oversampling factor - * @default 1 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - oversampling?: number; + id: string; /** - * Offset C - * @description Offset Cyan halfway between dots + * Is Intermediate + * @description Whether or not this is an intermediate invocation. * @default false */ - offset_c?: boolean; + is_intermediate?: boolean; /** - * Offset M - * @description Offset Magenta halfway between dots - * @default false + * Use Cache + * @description Whether or not to use the cache + * @default true */ - offset_m?: boolean; + use_cache?: boolean; /** - * Offset Y - * @description Offset Yellow halfway between dots - * @default false + * Image Width + * @description The image width, in pixels, to calculate tiles for. + * @default 1024 */ - offset_y?: boolean; + image_width?: number; /** - * Offset K - * @description Offset K halfway between dots - * @default false + * Image Height + * @description The image height, in pixels, to calculate tiles for. + * @default 1024 + */ + image_height?: number; + /** + * Tile Width + * @description The tile width, in pixels. + * @default 576 + */ + tile_width?: number; + /** + * Tile Height + * @description The tile height, in pixels. + * @default 576 + */ + tile_height?: number; + /** + * Overlap + * @description The target overlap, in pixels, between adjacent tiles. Adjacent tiles will overlap by at least this amount + * @default 128 */ - offset_k?: boolean; + overlap?: number; /** * type - * @default cmyk_halftone + * @default calculate_image_tiles * @constant */ - type: "cmyk_halftone"; + type: "calculate_image_tiles"; }; /** - * CMYK Merge - * @description Merge subtractive color channels (CMYK+alpha) + * Calculate Image Tiles Minimum Overlap + * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. */ - CMYKMergeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + CalculateImageTilesMinimumOverlapInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -4961,107 +4334,94 @@ export type components = { */ use_cache?: boolean; /** - * @description The c channel - * @default null - */ - c_channel?: components["schemas"]["ImageField"] | null; - /** - * @description The m channel - * @default null + * Image Width + * @description The image width, in pixels, to calculate tiles for. + * @default 1024 */ - m_channel?: components["schemas"]["ImageField"] | null; + image_width?: number; /** - * @description The y channel - * @default null + * Image Height + * @description The image height, in pixels, to calculate tiles for. + * @default 1024 */ - y_channel?: components["schemas"]["ImageField"] | null; + image_height?: number; /** - * @description The k channel - * @default null + * Tile Width + * @description The tile width, in pixels. + * @default 576 */ - k_channel?: components["schemas"]["ImageField"] | null; + tile_width?: number; /** - * @description The alpha channel - * @default null + * Tile Height + * @description The tile height, in pixels. + * @default 576 */ - alpha_channel?: components["schemas"]["ImageField"] | null; + tile_height?: number; /** - * Profile - * @description CMYK Color Profile - * @default Default - * @enum {string} + * Min Overlap + * @description Minimum overlap between adjacent tiles, in pixels. + * @default 128 */ - profile?: "Default" | "PIL"; + min_overlap?: number; /** * type - * @default cmyk_merge + * @default calculate_image_tiles_min_overlap * @constant */ - type: "cmyk_merge"; + type: "calculate_image_tiles_min_overlap"; }; - /** - * CMYKSeparationOutput - * @description Base class for invocations that output four L-mode images (C, M, Y, K) - */ - CMYKSeparationOutput: { - /** @description Blank image of the specified color */ - color_image: components["schemas"]["ImageField"]; - /** - * Width - * @description The width of the image in pixels - */ - width: number; - /** - * Height - * @description The height of the image in pixels - */ - height: number; - /** @description Blank image of the first separated color */ - part_a: components["schemas"]["ImageField"]; - /** - * Rgb Red A - * @description R value of color part A - */ - rgb_red_a: number; - /** - * Rgb Green A - * @description G value of color part A - */ - rgb_green_a: number; + /** CalculateImageTilesOutput */ + CalculateImageTilesOutput: { /** - * Rgb Blue A - * @description B value of color part A + * Tiles + * @description The tiles coordinates that cover a particular image shape. */ - rgb_blue_a: number; - /** @description Blank image of the second separated color */ - part_b: components["schemas"]["ImageField"]; + tiles: components["schemas"]["Tile"][]; /** - * Rgb Red B - * @description R value of color part B + * type + * @default calculate_image_tiles_output + * @constant */ - rgb_red_b: number; + type: "calculate_image_tiles_output"; + }; + /** + * CancelAllExceptCurrentResult + * @description Result of canceling all except current + */ + CancelAllExceptCurrentResult: { /** - * Rgb Green B - * @description G value of color part B + * Canceled + * @description Number of queue items canceled */ - rgb_green_b: number; + canceled: number; + }; + /** + * CancelByBatchIDsResult + * @description Result of canceling by list of batch ids + */ + CancelByBatchIDsResult: { /** - * Rgb Blue B - * @description B value of color part B + * Canceled + * @description Number of queue items canceled */ - rgb_blue_b: number; + canceled: number; + }; + /** + * CancelByDestinationResult + * @description Result of canceling by a destination + */ + CancelByDestinationResult: { /** - * type - * @default cmyk_separation_output - * @constant + * Canceled + * @description Number of queue items canceled */ - type: "cmyk_separation_output"; + canceled: number; }; /** - * CMYK Split - * @description Split an image into subtractive color channels (CMYK+alpha) + * Canny Edge Detection + * @description Geneartes an edge map using a cv2's Canny algorithm. */ - CMYKSplitInvocation: { + CannyEdgeDetectionInvocation: { /** * @description The board to save the image to * @default null @@ -5090,61 +4450,44 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to split into additive channels + * @description The image to process * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Profile - * @description CMYK Color Profile - * @default Default - * @enum {string} + * Low Threshold + * @description The low threshold of the Canny pixel gradient (0-255) + * @default 100 + */ + low_threshold?: number; + /** + * High Threshold + * @description The high threshold of the Canny pixel gradient (0-255) + * @default 200 */ - profile?: "Default" | "PIL"; + high_threshold?: number; /** * type - * @default cmyk_split + * @default canny_edge_detection * @constant */ - type: "cmyk_split"; + type: "canny_edge_detection"; }; /** - * CMYKSplitOutput - * @description Base class for invocations that output four L-mode images (C, M, Y, K) + * Canvas Paste Back + * @description Combines two images by using the mask provided. Intended for use on the Unified Canvas. */ - CMYKSplitOutput: { - /** @description Grayscale image of the cyan channel */ - c_channel: components["schemas"]["ImageField"]; - /** @description Grayscale image of the magenta channel */ - m_channel: components["schemas"]["ImageField"]; - /** @description Grayscale image of the yellow channel */ - y_channel: components["schemas"]["ImageField"]; - /** @description Grayscale image of the k channel */ - k_channel: components["schemas"]["ImageField"]; - /** @description Grayscale image of the alpha channel */ - alpha_channel: components["schemas"]["ImageField"]; + CanvasPasteBackInvocation: { /** - * Width - * @description The width of the image in pixels + * @description The board to save the image to + * @default null */ - width: number; - /** - * Height - * @description The height of the image in pixels - */ - height: number; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default cmyk_split_output - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "cmyk_split_output"; - }; - /** - * CSV To Index String - * @description CSVToIndexString converts a CSV to a String at index with a random option - */ - CSVToIndexStringInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5159,39 +4502,52 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Csv String - * @description csv string - * @default + * @description The source image + * @default null */ - csv_string?: string; + source_image?: components["schemas"]["ImageField"] | null; /** - * Random - * @description Random Index? - * @default true + * @description The target image + * @default null + */ + target_image?: components["schemas"]["ImageField"] | null; + /** + * @description The mask to use when pasting + * @default null */ - random?: boolean; + mask?: components["schemas"]["ImageField"] | null; /** - * Index - * @description zero based index into CSV array (note index will wrap around if out of bounds) + * Mask Blur + * @description The amount to blur the mask by * @default 0 */ - index?: number; + mask_blur?: number; /** * type - * @default csv_to_index_string + * @default canvas_paste_back * @constant */ - type: "csv_to_index_string"; + type: "canvas_paste_back"; }; /** - * CSV To Strings - * @description Converts a CSV string to a collection of strings + * Canvas V2 Mask and Crop + * @description Handles Canvas V2 image output masking and cropping */ - CSVToStringsInvocation: { + CanvasV2MaskAndCropInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5210,33 +4566,38 @@ export type components = { */ use_cache?: boolean; /** - * Csv String - * @description csv string + * @description The source image onto which the masked generated image is pasted. If omitted, the masked generated image is returned with transparency. * @default null */ - csv_string?: string | null; + source_image?: components["schemas"]["ImageField"] | null; /** - * type - * @default csv_to_strings - * @constant + * @description The image to apply the mask to + * @default null */ - type: "csv_to_strings"; - }; - /** - * CV2 Infill - * @description Infills transparent areas of an image using OpenCV Inpainting - */ - CV2InfillInvocation: { + generated_image?: components["schemas"]["ImageField"] | null; /** - * @description The board to save the image to + * @description The mask to apply * @default null */ - board?: components["schemas"]["BoardField"] | null; + mask?: components["schemas"]["ImageField"] | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Mask Blur + * @description The amount to blur the mask by + * @default 0 */ - metadata?: components["schemas"]["MetadataField"] | null; + mask_blur?: number; + /** + * type + * @default canvas_v2_mask_and_crop + * @constant + */ + type: "canvas_v2_mask_and_crop"; + }; + /** + * Center Pad or Crop Image + * @description Pad or crop an image's sides from the center by specified pixels. Positive values are outside of the image. + */ + CenterPadCropInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5255,59 +4616,110 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The image to crop * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * type - * @default infill_cv2 - * @constant + * Left + * @description Number of pixels to pad/crop from the left (negative values crop inwards, positive values pad outwards) + * @default 0 */ - type: "infill_cv2"; - }; - /** CacheStats */ - CacheStats: { + left?: number; /** - * Hits + * Right + * @description Number of pixels to pad/crop from the right (negative values crop inwards, positive values pad outwards) * @default 0 */ - hits?: number; + right?: number; /** - * Misses + * Top + * @description Number of pixels to pad/crop from the top (negative values crop inwards, positive values pad outwards) * @default 0 */ - misses?: number; + top?: number; /** - * High Watermark + * Bottom + * @description Number of pixels to pad/crop from the bottom (negative values crop inwards, positive values pad outwards) * @default 0 */ - high_watermark?: number; + bottom?: number; /** - * In Cache - * @default 0 + * type + * @default img_pad_crop + * @constant */ - in_cache?: number; + type: "img_pad_crop"; + }; + /** + * Classification + * @description The classification of an Invocation. + * - `Stable`: The invocation, including its inputs/outputs and internal logic, is stable. You may build workflows with it, having confidence that they will not break because of a change in this invocation. + * - `Beta`: The invocation is not yet stable, but is planned to be stable in the future. Workflows built around this invocation may break, but we are committed to supporting this invocation long-term. + * - `Prototype`: The invocation is not yet stable and may be removed from the application at any time. Workflows built around this invocation may break, and we are *not* committed to supporting this invocation. + * - `Deprecated`: The invocation is deprecated and may be removed in a future version. + * - `Internal`: The invocation is not intended for use by end-users. It may be changed or removed at any time, but is exposed for users to play with. + * - `Special`: The invocation is a special case and does not fit into any of the other classifications. + * @enum {string} + */ + Classification: "stable" | "beta" | "prototype" | "deprecated" | "internal" | "special"; + /** + * ClearResult + * @description Result of clearing the session queue + */ + ClearResult: { /** - * Cleared - * @default 0 + * Deleted + * @description Number of queue items deleted */ - cleared?: number; + deleted: number; + }; + /** + * ClipVariantType + * @description Variant type. + * @enum {string} + */ + ClipVariantType: "large" | "gigantic"; + /** + * CogView4ConditioningField + * @description A conditioning tensor primitive value + */ + CogView4ConditioningField: { /** - * Cache Size - * @default 0 + * Conditioning Name + * @description The name of conditioning tensor */ - cache_size?: number; - /** Loaded Model Sizes */ - loaded_model_sizes?: { - [key: string]: number; - }; + conditioning_name: string; }; /** - * Calculate Image Tiles Even Split - * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. + * CogView4ConditioningOutput + * @description Base class for nodes that output a CogView text conditioning tensor. */ - CalculateImageTilesEvenSplitInvocation: { + CogView4ConditioningOutput: { + /** @description Conditioning tensor */ + conditioning: components["schemas"]["CogView4ConditioningField"]; + /** + * type + * @default cogview4_conditioning_output + * @constant + */ + type: "cogview4_conditioning_output"; + }; + /** + * Denoise - CogView4 + * @description Run the denoising process with a CogView4 model. + */ + CogView4DenoiseInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5326,106 +4738,95 @@ export type components = { */ use_cache?: boolean; /** - * Image Width - * @description The image width, in pixels, to calculate tiles for. - * @default 1024 + * @description Latents tensor + * @default null */ - image_width?: number; + latents?: components["schemas"]["LatentsField"] | null; /** - * Image Height - * @description The image height, in pixels, to calculate tiles for. - * @default 1024 + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - image_height?: number; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Num Tiles X - * @description Number of tiles to divide image into on the x axis - * @default 2 + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - num_tiles_x?: number; + denoising_start?: number; /** - * Num Tiles Y - * @description Number of tiles to divide image into on the y axis - * @default 2 + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - num_tiles_y?: number; + denoising_end?: number; /** - * Overlap - * @description The overlap, in pixels, between adjacent tiles. - * @default 128 + * Transformer + * @description CogView4 model (Transformer) to load + * @default null */ - overlap?: number; + transformer?: components["schemas"]["TransformerField"] | null; /** - * type - * @default calculate_image_tiles_even_split - * @constant + * @description Positive conditioning tensor + * @default null */ - type: "calculate_image_tiles_even_split"; - }; - /** - * Calculate Image Tiles - * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. - */ - CalculateImageTilesInvocation: { + positive_conditioning?: components["schemas"]["CogView4ConditioningField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description Negative conditioning tensor + * @default null */ - id: string; + negative_conditioning?: components["schemas"]["CogView4ConditioningField"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 3.5 */ - use_cache?: boolean; + cfg_scale?: number | number[]; /** - * Image Width - * @description The image width, in pixels, to calculate tiles for. + * Width + * @description Width of the generated image. * @default 1024 */ - image_width?: number; + width?: number; /** - * Image Height - * @description The image height, in pixels, to calculate tiles for. + * Height + * @description Height of the generated image. * @default 1024 */ - image_height?: number; - /** - * Tile Width - * @description The tile width, in pixels. - * @default 576 - */ - tile_width?: number; + height?: number; /** - * Tile Height - * @description The tile height, in pixels. - * @default 576 + * Steps + * @description Number of steps to run + * @default 25 */ - tile_height?: number; + steps?: number; /** - * Overlap - * @description The target overlap, in pixels, between adjacent tiles. Adjacent tiles will overlap by at least this amount - * @default 128 + * Seed + * @description Randomness seed for reproducibility. + * @default 0 */ - overlap?: number; + seed?: number; /** * type - * @default calculate_image_tiles + * @default cogview4_denoise * @constant */ - type: "calculate_image_tiles"; + type: "cogview4_denoise"; }; /** - * Calculate Image Tiles Minimum Overlap - * @description Calculate the coordinates and overlaps of tiles that cover a target image shape. + * Image to Latents - CogView4 + * @description Generates latents from an image. */ - CalculateImageTilesMinimumOverlapInvocation: { + CogView4ImageToLatentsInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5444,94 +4845,27 @@ export type components = { */ use_cache?: boolean; /** - * Image Width - * @description The image width, in pixels, to calculate tiles for. - * @default 1024 - */ - image_width?: number; - /** - * Image Height - * @description The image height, in pixels, to calculate tiles for. - * @default 1024 - */ - image_height?: number; - /** - * Tile Width - * @description The tile width, in pixels. - * @default 576 - */ - tile_width?: number; - /** - * Tile Height - * @description The tile height, in pixels. - * @default 576 - */ - tile_height?: number; - /** - * Min Overlap - * @description Minimum overlap between adjacent tiles, in pixels. - * @default 128 - */ - min_overlap?: number; - /** - * type - * @default calculate_image_tiles_min_overlap - * @constant + * @description The image to encode. + * @default null */ - type: "calculate_image_tiles_min_overlap"; - }; - /** CalculateImageTilesOutput */ - CalculateImageTilesOutput: { + image?: components["schemas"]["ImageField"] | null; /** - * Tiles - * @description The tiles coordinates that cover a particular image shape. + * @description VAE + * @default null */ - tiles: components["schemas"]["Tile"][]; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default calculate_image_tiles_output + * @default cogview4_i2l * @constant */ - type: "calculate_image_tiles_output"; - }; - /** - * CancelAllExceptCurrentResult - * @description Result of canceling all except current - */ - CancelAllExceptCurrentResult: { - /** - * Canceled - * @description Number of queue items canceled - */ - canceled: number; - }; - /** - * CancelByBatchIDsResult - * @description Result of canceling by list of batch ids - */ - CancelByBatchIDsResult: { - /** - * Canceled - * @description Number of queue items canceled - */ - canceled: number; - }; - /** - * CancelByDestinationResult - * @description Result of canceling by a destination - */ - CancelByDestinationResult: { - /** - * Canceled - * @description Number of queue items canceled - */ - canceled: number; + type: "cogview4_i2l"; }; /** - * Canny Edge Detection - * @description Geneartes an edge map using a cv2's Canny algorithm. + * Latents to Image - CogView4 + * @description Generates an image from latents. */ - CannyEdgeDetectionInvocation: { + CogView4LatentsToImageInvocation: { /** * @description The board to save the image to * @default null @@ -5560,44 +4894,27 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description Latents tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * Low Threshold - * @description The low threshold of the Canny pixel gradient (0-255) - * @default 100 - */ - low_threshold?: number; + latents?: components["schemas"]["LatentsField"] | null; /** - * High Threshold - * @description The high threshold of the Canny pixel gradient (0-255) - * @default 200 + * @description VAE + * @default null */ - high_threshold?: number; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default canny_edge_detection + * @default cogview4_l2i * @constant */ - type: "canny_edge_detection"; + type: "cogview4_l2i"; }; /** - * Canvas Paste Back - * @description Combines two images by using the mask provided. Intended for use on the Unified Canvas. + * Main Model - CogView4 + * @description Loads a CogView4 base model, outputting its submodels. */ - CanvasPasteBackInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + CogView4ModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5615,49 +4932,47 @@ export type components = { * @default true */ use_cache?: boolean; + /** @description CogView4 model (Transformer) to load */ + model: components["schemas"]["ModelIdentifierField"]; /** - * @description The source image - * @default null + * type + * @default cogview4_model_loader + * @constant */ - source_image?: components["schemas"]["ImageField"] | null; + type: "cogview4_model_loader"; + }; + /** + * CogView4ModelLoaderOutput + * @description CogView4 base model loader output. + */ + CogView4ModelLoaderOutput: { /** - * @description The target image - * @default null + * Transformer + * @description Transformer */ - target_image?: components["schemas"]["ImageField"] | null; + transformer: components["schemas"]["TransformerField"]; /** - * @description The mask to use when pasting - * @default null + * GLM Encoder + * @description GLM (THUDM) tokenizer and text encoder */ - mask?: components["schemas"]["ImageField"] | null; + glm_encoder: components["schemas"]["GlmEncoderField"]; /** - * Mask Blur - * @description The amount to blur the mask by - * @default 0 + * VAE + * @description VAE */ - mask_blur?: number; + vae: components["schemas"]["VAEField"]; /** * type - * @default canvas_paste_back + * @default cogview4_model_loader_output * @constant */ - type: "canvas_paste_back"; + type: "cogview4_model_loader_output"; }; /** - * Canvas V2 Mask and Crop - * @description Handles Canvas V2 image output masking and cropping + * Prompt - CogView4 + * @description Encodes and preps a prompt for a cogview4 image. */ - CanvasV2MaskAndCropInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + CogView4TextEncoderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5676,38 +4991,29 @@ export type components = { */ use_cache?: boolean; /** - * @description The source image onto which the masked generated image is pasted. If omitted, the masked generated image is returned with transparency. + * Prompt + * @description Text prompt to encode. * @default null */ - source_image?: components["schemas"]["ImageField"] | null; + prompt?: string | null; /** - * @description The image to apply the mask to + * GLM Encoder + * @description GLM (THUDM) tokenizer and text encoder * @default null */ - generated_image?: components["schemas"]["ImageField"] | null; - /** - * @description The mask to apply - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Mask Blur - * @description The amount to blur the mask by - * @default 0 - */ - mask_blur?: number; + glm_encoder?: components["schemas"]["GlmEncoderField"] | null; /** * type - * @default canvas_v2_mask_and_crop + * @default cogview4_text_encoder * @constant */ - type: "canvas_v2_mask_and_crop"; + type: "cogview4_text_encoder"; }; /** - * Center Pad or Crop Image - * @description Pad or crop an image's sides from the center by specified pixels. Positive values are outside of the image. + * CollectInvocation + * @description Collects values into a collection */ - CenterPadCropInvocation: { + CollectInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5726,46 +5032,61 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to crop + * Collection Item + * @description The item to collect (all inputs must be of the same type) * @default null */ - image?: components["schemas"]["ImageField"] | null; + item?: unknown | null; /** - * Left - * @description Number of pixels to pad/crop from the left (negative values crop inwards, positive values pad outwards) - * @default 0 + * Collection + * @description An optional collection to append to + * @default [] */ - left?: number; + collection?: unknown[]; /** - * Right - * @description Number of pixels to pad/crop from the right (negative values crop inwards, positive values pad outwards) - * @default 0 + * type + * @default collect + * @constant */ - right?: number; + type: "collect"; + }; + /** CollectInvocationOutput */ + CollectInvocationOutput: { /** - * Top - * @description Number of pixels to pad/crop from the top (negative values crop inwards, positive values pad outwards) - * @default 0 + * Collection + * @description The collection of input items */ - top?: number; + collection: unknown[]; /** - * Bottom - * @description Number of pixels to pad/crop from the bottom (negative values crop inwards, positive values pad outwards) - * @default 0 + * type + * @default collect_output + * @constant */ - bottom?: number; + type: "collect_output"; + }; + /** + * ColorCollectionOutput + * @description Base class for nodes that output a collection of colors + */ + ColorCollectionOutput: { + /** + * Collection + * @description The output colors + */ + collection: components["schemas"]["ColorField"][]; /** * type - * @default img_pad_crop + * @default color_collection_output * @constant */ - type: "img_pad_crop"; + type: "color_collection_output"; }; /** - * ChromaNoise - * @description Adds chroma-only noise (in OKLab space) to an image. + * Color Correct + * @description Matches the color histogram of a base image to a reference image, optionally + * using a mask to only color-correct certain regions of the base image. */ - ChromaNoiseInvocation: { + ColorCorrectInvocation: { /** * @description The board to save the image to * @default null @@ -5794,68 +5115,65 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to add chroma noise to + * @description The image to color-correct * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * Amount 1 - * @description Amount of the first chroma noise layer - * @default 100 - */ - amount_1?: number; - /** - * Amount 2 - * @description Amount of the second chroma noise layer - * @default 50 - */ - amount_2?: number; + base_image?: components["schemas"]["ImageField"] | null; /** - * Seed 1 - * @description The first seed to use (omit for random) + * @description Reference image for color-correction * @default null */ - seed_1?: number | null; + color_reference?: components["schemas"]["ImageField"] | null; /** - * Seed 2 - * @description The second seed to use (omit for random) + * @description Optional mask to limit color correction area * @default null */ - seed_2?: number | null; - /** - * Blur 1 - * @description The strength of the first noise blur - * @default 0.5 - */ - blur_1?: number; + mask?: components["schemas"]["ImageField"] | null; /** - * Blur 2 - * @description The strength of the second noise blur - * @default 0.5 + * Color Space + * @description Colorspace in which to apply histogram matching + * @default RGB + * @enum {string} */ - blur_2?: number; + colorspace?: "RGB" | "YCbCr" | "YCbCr-Chroma" | "YCbCr-Luma"; /** * type - * @default chroma_noise + * @default color_correct * @constant */ - type: "chroma_noise"; + type: "color_correct"; }; /** - * Chromatic Aberration - * @description Simulate realistic chromatic aberration in an image with controllable strength and center point. + * ColorField + * @description A color primitive field */ - ChromaticAberrationInvocation: { + ColorField: { /** - * @description Optional metadata to be saved with the image - * @default null + * R + * @description The red component */ - metadata?: components["schemas"]["MetadataField"] | null; + r: number; /** - * @description The board to save the image to - * @default null + * G + * @description The green component */ - board?: components["schemas"]["BoardField"] | null; + g: number; + /** + * B + * @description The blue component + */ + b: number; + /** + * A + * @description The alpha component + */ + a: number; + }; + /** + * Color Primitive + * @description A color primitive value + */ + ColorInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -5874,71 +5192,27 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to apply chromatic aberration to - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Strength - * @description Strength of the chromatic shift (0–10) - * @default 1 - */ - strength?: number; - /** - * Center X - * @description X coordinate of center (0–1) - * @default 0.5 - */ - center_x?: number; - /** - * Center Y - * @description Y coordinate of center (0–1) - * @default 0.5 + * @description The color value + * @default { + * "r": 0, + * "g": 0, + * "b": 0, + * "a": 255 + * } */ - center_y?: number; + color?: components["schemas"]["ColorField"]; /** * type - * @default chromatic_aberration + * @default color * @constant */ - type: "chromatic_aberration"; - }; - /** - * Classification - * @description The classification of an Invocation. - * - `Stable`: The invocation, including its inputs/outputs and internal logic, is stable. You may build workflows with it, having confidence that they will not break because of a change in this invocation. - * - `Beta`: The invocation is not yet stable, but is planned to be stable in the future. Workflows built around this invocation may break, but we are committed to supporting this invocation long-term. - * - `Prototype`: The invocation is not yet stable and may be removed from the application at any time. Workflows built around this invocation may break, and we are *not* committed to supporting this invocation. - * - `Deprecated`: The invocation is deprecated and may be removed in a future version. - * - `Internal`: The invocation is not intended for use by end-users. It may be changed or removed at any time, but is exposed for users to play with. - * - `Special`: The invocation is a special case and does not fit into any of the other classifications. - * @enum {string} - */ - Classification: "stable" | "beta" | "prototype" | "deprecated" | "internal" | "special"; - /** - * ClearResult - * @description Result of clearing the session queue - */ - ClearResult: { - /** - * Deleted - * @description Number of queue items deleted - */ - deleted: number; + type: "color"; }; /** - * ClipVariantType - * @description Variant type. - * @enum {string} - */ - ClipVariantType: "large" | "gigantic"; - /** - * Clipseg Mask Hierarchy - * @description Creates a segmentation hierarchy of mutually exclusive masks from clipseg text prompts. - * - * This node takes up to seven pairs of prompts/threshold values, then descends through them hierarchically creating mutually exclusive masks out of whatever it can match from the input image. This means whatever is matched in prompt 1 will be subtracted from the match area for prompt 2; both areas will be omitted from the match area of prompt 3; etc. The idea is that by starting with foreground objects and working your way back through a scene, you can create a more-or-less complete segmentation map for the image whose constituent segments can be passed off to different masks for regional conditioning or other processing. + * Color Map + * @description Generates a color map from the provided image. */ - ClipsegMaskHierarchyInvocation: { + ColorMapInvocation: { /** * @description The board to save the image to * @default null @@ -5967,181 +5241,88 @@ export type components = { */ use_cache?: boolean; /** - * @description The image from which to create masks + * @description The image to process * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Invert Output - * @description Off: white on black / On: black on white - * @default true - */ - invert_output?: boolean; - /** - * Smoothing - * @description Radius of blur to apply before thresholding - * @default 4 - */ - smoothing?: number; - /** - * Prompt 1 - * @description Text to mask prompt with highest segmentation priority - * @default null - */ - prompt_1?: string | null; - /** - * Threshold 1 - * @description Detection confidence threshold for prompt 1 - * @default 0.4 + * Tile Size + * @description Tile size + * @default 64 */ - threshold_1?: number; + tile_size?: number; /** - * Prompt 2 - * @description Text to mask prompt, behind prompt 1 - * @default null + * type + * @default color_map + * @constant */ - prompt_2?: string | null; + type: "color_map"; + }; + /** + * ColorOutput + * @description Base class for nodes that output a single color + */ + ColorOutput: { + /** @description The output color */ + color: components["schemas"]["ColorField"]; /** - * Threshold 2 - * @description Detection confidence threshold for prompt 2 - * @default 0.4 + * type + * @default color_output + * @constant */ - threshold_2?: number; + type: "color_output"; + }; + /** + * Prompt - SD1.5 + * @description Parse prompt using compel package to conditioning. + */ + CompelInvocation: { /** - * Prompt 3 - * @description Text to mask prompt, behind prompts 1 & 2 - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - prompt_3?: string | null; + id: string; /** - * Threshold 3 - * @description Detection confidence threshold for prompt 3 - * @default 0.4 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - threshold_3?: number; + is_intermediate?: boolean; /** - * Prompt 4 - * @description Text to mask prompt, behind prompts 1, 2, & 3 - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - prompt_4?: string | null; + use_cache?: boolean; /** - * Threshold 4 - * @description Detection confidence threshold for prompt 4 - * @default 0.4 + * Prompt + * @description Prompt to be parsed by Compel to create a conditioning tensor + * @default */ - threshold_4?: number; + prompt?: string; /** - * Prompt 5 - * @description Text to mask prompt, behind prompts 1 thru 4 + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - prompt_5?: string | null; - /** - * Threshold 5 - * @description Detection confidence threshold for prompt 5 - * @default 0.4 - */ - threshold_5?: number; + clip?: components["schemas"]["CLIPField"] | null; /** - * Prompt 6 - * @description Text to mask prompt, behind prompts 1 thru 5 + * @description A mask defining the region that this conditioning prompt applies to. * @default null */ - prompt_6?: string | null; + mask?: components["schemas"]["TensorField"] | null; /** - * Threshold 6 - * @description Detection confidence threshold for prompt 6 - * @default 0.4 + * type + * @default compel + * @constant */ - threshold_6?: number; - /** - * Prompt 7 - * @description Text to mask prompt, lowest priority behind all others - * @default null - */ - prompt_7?: string | null; - /** - * Threshold 7 - * @description Detection confidence threshold for prompt 7 - * @default 0.4 - */ - threshold_7?: number; - /** - * type - * @default clipseg_mask_hierarchy - * @constant - */ - type: "clipseg_mask_hierarchy"; - }; - /** - * ClipsegMaskHierarchyOutput - * @description Class for invocations that output a hierarchy of masks - */ - ClipsegMaskHierarchyOutput: { - /** @description Mask corresponding to prompt 1 (full coverage) */ - mask_1: components["schemas"]["ImageField"]; - /** @description Mask corresponding to prompt 2 (minus mask 1) */ - mask_2: components["schemas"]["ImageField"]; - /** @description Mask corresponding to prompt 3 (minus masks 1 & 2) */ - mask_3: components["schemas"]["ImageField"]; - /** @description Mask corresponding to prompt 4 (minus masks 1, 2, & 3) */ - mask_4: components["schemas"]["ImageField"]; - /** @description Mask corresponding to prompt 5 (minus masks 1 thru 4) */ - mask_5: components["schemas"]["ImageField"]; - /** @description Mask corresponding to prompt 6 (minus masks 1 thru 5) */ - mask_6: components["schemas"]["ImageField"]; - /** @description Mask corresponding to prompt 7 (minus masks 1 thru 6) */ - mask_7: components["schemas"]["ImageField"]; - /** @description Mask coresponding to remaining unmatched image areas. */ - ground_mask: components["schemas"]["ImageField"]; - /** - * type - * @default clipseg_mask_hierarchy_output - * @constant - */ - type: "clipseg_mask_hierarchy_output"; - }; - /** - * CogView4ConditioningField - * @description A conditioning tensor primitive value - */ - CogView4ConditioningField: { - /** - * Conditioning Name - * @description The name of conditioning tensor - */ - conditioning_name: string; - }; - /** - * CogView4ConditioningOutput - * @description Base class for nodes that output a CogView text conditioning tensor. - */ - CogView4ConditioningOutput: { - /** @description Conditioning tensor */ - conditioning: components["schemas"]["CogView4ConditioningField"]; - /** - * type - * @default cogview4_conditioning_output - * @constant - */ - type: "cogview4_conditioning_output"; + type: "compel"; }; /** - * Denoise - CogView4 - * @description Run the denoising process with a CogView4 model. + * Conditioning Collection Primitive + * @description A collection of conditioning tensor primitive values */ - CogView4DenoiseInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + ConditioningCollectionInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -6160,95 +5341,56 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. - * @default null - */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; - /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 - */ - denoising_start?: number; - /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 - */ - denoising_end?: number; - /** - * Transformer - * @description CogView4 model (Transformer) to load - * @default null - */ - transformer?: components["schemas"]["TransformerField"] | null; - /** - * @description Positive conditioning tensor - * @default null - */ - positive_conditioning?: components["schemas"]["CogView4ConditioningField"] | null; - /** - * @description Negative conditioning tensor - * @default null - */ - negative_conditioning?: components["schemas"]["CogView4ConditioningField"] | null; - /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 3.5 - */ - cfg_scale?: number | number[]; - /** - * Width - * @description Width of the generated image. - * @default 1024 - */ - width?: number; - /** - * Height - * @description Height of the generated image. - * @default 1024 + * Collection + * @description The collection of conditioning tensors + * @default [] */ - height?: number; + collection?: components["schemas"]["ConditioningField"][]; /** - * Steps - * @description Number of steps to run - * @default 25 + * type + * @default conditioning_collection + * @constant */ - steps?: number; + type: "conditioning_collection"; + }; + /** + * ConditioningCollectionOutput + * @description Base class for nodes that output a collection of conditioning tensors + */ + ConditioningCollectionOutput: { /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * Collection + * @description The output conditioning tensors */ - seed?: number; + collection: components["schemas"]["ConditioningField"][]; /** * type - * @default cogview4_denoise + * @default conditioning_collection_output * @constant */ - type: "cogview4_denoise"; + type: "conditioning_collection_output"; }; /** - * Image to Latents - CogView4 - * @description Generates latents from an image. + * ConditioningField + * @description A conditioning tensor primitive value */ - CogView4ImageToLatentsInvocation: { + ConditioningField: { /** - * @description The board to save the image to - * @default null + * Conditioning Name + * @description The name of conditioning tensor */ - board?: components["schemas"]["BoardField"] | null; + conditioning_name: string; /** - * @description Optional metadata to be saved with the image + * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + mask?: components["schemas"]["TensorField"] | null; + }; + /** + * Conditioning Primitive + * @description A conditioning tensor primitive value + */ + ConditioningInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -6267,27 +5409,36 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to encode. + * @description Conditioning tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; + conditioning?: components["schemas"]["ConditioningField"] | null; /** - * @description VAE - * @default null + * type + * @default conditioning + * @constant */ - vae?: components["schemas"]["VAEField"] | null; + type: "conditioning"; + }; + /** + * ConditioningOutput + * @description Base class for nodes that output a single conditioning tensor + */ + ConditioningOutput: { + /** @description Conditioning tensor */ + conditioning: components["schemas"]["ConditioningField"]; /** * type - * @default cogview4_i2l + * @default conditioning_output * @constant */ - type: "cogview4_i2l"; + type: "conditioning_output"; }; /** - * Latents to Image - CogView4 - * @description Generates an image from latents. + * Content Shuffle + * @description Shuffles the image, similar to a 'liquify' filter. */ - CogView4LatentsToImageInvocation: { + ContentShuffleInvocation: { /** * @description The board to save the image to * @default null @@ -6316,181 +5467,158 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * @description The image to process * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description VAE - * @default null + * Scale Factor + * @description The scale factor used for the shuffle + * @default 256 */ - vae?: components["schemas"]["VAEField"] | null; + scale_factor?: number; /** * type - * @default cogview4_l2i + * @default content_shuffle * @constant */ - type: "cogview4_l2i"; + type: "content_shuffle"; }; - /** - * Main Model - CogView4 - * @description Loads a CogView4 base model, outputting its submodels. - */ - CogView4ModelLoaderInvocation: { + /** ControlAdapterDefaultSettings */ + ControlAdapterDefaultSettings: { + /** Preprocessor */ + preprocessor: string | null; + }; + /** ControlField */ + ControlField: { + /** @description The control image */ + image: components["schemas"]["ImageField"]; + /** @description The ControlNet model to use */ + control_model: components["schemas"]["ModelIdentifierField"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - id: string; + control_weight?: number | number[]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - is_intermediate?: boolean; + begin_step_percent?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - use_cache?: boolean; - /** @description CogView4 model (Transformer) to load */ - model: components["schemas"]["ModelIdentifierField"]; + end_step_percent?: number; /** - * type - * @default cogview4_model_loader - * @constant + * Control Mode + * @description The control mode to use + * @default balanced + * @enum {string} */ - type: "cogview4_model_loader"; - }; - /** - * CogView4ModelLoaderOutput - * @description CogView4 base model loader output. - */ - CogView4ModelLoaderOutput: { - /** - * Transformer - * @description Transformer - */ - transformer: components["schemas"]["TransformerField"]; - /** - * GLM Encoder - * @description GLM (THUDM) tokenizer and text encoder - */ - glm_encoder: components["schemas"]["GlmEncoderField"]; + control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; /** - * VAE - * @description VAE + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} */ - vae: components["schemas"]["VAEField"]; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + }; + /** ControlLoRAField */ + ControlLoRAField: { + /** @description Info to load lora model */ + lora: components["schemas"]["ModelIdentifierField"]; /** - * type - * @default cogview4_model_loader_output - * @constant + * Weight + * @description Weight to apply to lora model */ - type: "cogview4_model_loader_output"; + weight: number; + /** @description Image to use in structural conditioning */ + img: components["schemas"]["ImageField"]; }; /** - * Prompt - CogView4 - * @description Encodes and preps a prompt for a cogview4 image. + * ControlLoRA_LyCORIS_FLUX_Config + * @description Model config for Control LoRA models. */ - CogView4TextEncoderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + ControlLoRA_LyCORIS_FLUX_Config: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Key + * @description A unique key for this model. */ - use_cache?: boolean; + key: string; /** - * Prompt - * @description Text prompt to encode. - * @default null + * Hash + * @description The hash of the model file(s). */ - prompt?: string | null; + hash: string; /** - * GLM Encoder - * @description GLM (THUDM) tokenizer and text encoder - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - glm_encoder?: components["schemas"]["GlmEncoderField"] | null; + path: string; /** - * type - * @default cogview4_text_encoder - * @constant + * File Size + * @description The size of the model in bytes. */ - type: "cogview4_text_encoder"; - }; - /** - * CollectInvocation - * @description Collects values into a collection - */ - CollectInvocation: { + file_size: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Name + * @description Name of the model. */ - id: string; + name: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Description + * @description Model description */ - is_intermediate?: boolean; + description: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source + * @description The original source of the model (path, URL or repo_id). */ - use_cache?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Collection Item - * @description The item to collect (all inputs must be of the same type) - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - item?: unknown | null; + source_api_response: string | null; /** - * Collection - * @description An optional collection to append to - * @default [] + * Cover Image + * @description Url for image to preview model */ - collection?: unknown[]; + cover_image: string | null; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * type - * @default collect + * Base + * @default flux * @constant */ - type: "collect"; - }; - /** CollectInvocationOutput */ - CollectInvocationOutput: { + base: "flux"; /** - * Collection - * @description The collection of input items + * Type + * @default control_lora + * @constant */ - collection: unknown[]; + type: "control_lora"; /** - * type - * @default collect_output + * Format + * @default lycoris * @constant */ - type: "collect_output"; + format: "lycoris"; + /** Trigger Phrases */ + trigger_phrases: string[] | null; }; /** - * Collection Count - * @description Counts the number of items in a collection. + * ControlNet - SD1.5, SD2, SDXL + * @description Collects ControlNet info to pass to other nodes */ - CollectionCountInvocation: { + ControlNetInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -6505,1001 +5633,822 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Collection - * @description The collection to count - * @default [] - */ - collection?: unknown[]; - /** - * type - * @default collection_count - * @constant - */ - type: "collection_count"; - }; - /** - * CollectionCountOutput - * @description The output of the collection count node. - */ - CollectionCountOutput: { - /** - * Count - * @description The number of items in the collection - */ - count: number; - /** - * type - * @default collection_count_output - * @constant + * @description The control image + * @default null */ - type: "collection_count_output"; - }; - /** - * Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option - */ - CollectionIndexInvocation: { + image?: components["schemas"]["ImageField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description ControlNet model to load + * @default null */ - id: string; + control_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - is_intermediate?: boolean; + control_weight?: number | number[]; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - use_cache?: boolean; + begin_step_percent?: number; /** - * Random - * @description Random Index? - * @default true + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - random?: boolean; + end_step_percent?: number; /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) - * @default 0 + * Control Mode + * @description The control mode used + * @default balanced + * @enum {string} */ - index?: number; + control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; /** - * Collection - * @description collection - * @default null + * Resize Mode + * @description The resize mode used + * @default just_resize + * @enum {string} */ - collection?: unknown[] | null; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** * type - * @default collection_index + * @default controlnet * @constant */ - type: "collection_index"; + type: "controlnet"; }; - /** - * CollectionIndexOutput - * @description Used to connect iteration outputs. Will be expanded to a specific output. - */ - CollectionIndexOutput: { + /** ControlNetMetadataField */ + ControlNetMetadataField: { + /** @description The control image */ + image: components["schemas"]["ImageField"]; /** - * Collection Item - * @description The item being iterated over + * @description The control image, after processing. + * @default null */ - item: unknown; + processed_image?: components["schemas"]["ImageField"] | null; + /** @description The ControlNet model to use */ + control_model: components["schemas"]["ModelIdentifierField"]; /** - * Index - * @description The index of the selected item + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - index: number; + control_weight?: number | number[]; /** - * Total - * @description The total number of items in the collection + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - total: number; + begin_step_percent?: number; /** - * type - * @default collection_index_output - * @constant + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - type: "collection_index_output"; - }; - /** - * Collection Join - * @description CollectionJoin Joins two collections into a single collection - */ - CollectionJoinInvocation: { + end_step_percent?: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Control Mode + * @description The control mode to use + * @default balanced + * @enum {string} */ - id: string; + control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} */ - is_intermediate?: boolean; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + }; + /** + * ControlNetRecallParameter + * @description ControlNet configuration for recall + */ + ControlNetRecallParameter: { /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Model Name + * @description The name of the ControlNet/T2I Adapter/Control LoRA model */ - use_cache?: boolean; + model_name: string; /** - * Collection A - * @description collection - * @default [] + * Image Name + * @description The filename of the control image in outputs/images */ - collection_a?: unknown[]; + image_name?: string | null; /** - * Collection B - * @description collection - * @default [] + * Weight + * @description The weight for the control adapter + * @default 1 */ - collection_b?: unknown[]; + weight?: number; /** - * type - * @default collection_join - * @constant + * Begin Step Percent + * @description When the control adapter is first applied (% of total steps) */ - type: "collection_join"; - }; - /** CollectionJoinOutput */ - CollectionJoinOutput: { + begin_step_percent?: number | null; /** - * Collection - * @description The collection of output items + * End Step Percent + * @description When the control adapter is last applied (% of total steps) */ - collection: unknown[]; + end_step_percent?: number | null; /** - * type - * @default collection_join_output - * @constant + * Control Mode + * @description The control mode (ControlNet only) */ - type: "collection_join_output"; + control_mode?: ("balanced" | "more_prompt" | "more_control") | null; }; - /** - * Collection Reverse - * @description Reverses a collection. - */ - CollectionReverseInvocation: { + /** ControlNet_Checkpoint_FLUX_Config */ + ControlNet_Checkpoint_FLUX_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Collection - * @description The collection to reverse - * @default [] + * File Size + * @description The size of the model in bytes. */ - collection?: unknown[]; + file_size: number; /** - * type - * @default collection_reverse - * @constant + * Name + * @description Name of the model. */ - type: "collection_reverse"; - }; - /** - * CollectionReverseOutput - * @description The output of the collection reverse node. - */ - CollectionReverseOutput: { + name: string; /** - * Collection - * @description The reversed collection + * Description + * @description Model description */ - collection: unknown[]; + description: string | null; /** - * type - * @default collection_reverse_output - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "collection_reverse_output"; - }; - /** - * Collection Slice - * @description Slices a collection. - */ - CollectionSliceInvocation: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - id: string; + source_api_response: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Cover Image + * @description Url for image to preview model */ - is_intermediate?: boolean; + cover_image: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Config Path + * @description Path to the config for this model, if any. */ - use_cache?: boolean; + config_path: string | null; /** - * Collection - * @description The collection to slice - * @default [] + * Type + * @default controlnet + * @constant */ - collection?: unknown[]; + type: "controlnet"; /** - * Start - * @description The start index of the slice - * @default 0 + * Format + * @default checkpoint + * @constant */ - start?: number; + format: "checkpoint"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Stop - * @description The stop index of the slice (exclusive) - * @default null + * Base + * @default flux + * @constant */ - stop?: number | null; + base: "flux"; + }; + /** ControlNet_Checkpoint_SD1_Config */ + ControlNet_Checkpoint_SD1_Config: { /** - * Step - * @description The step of the slice - * @default 1 + * Key + * @description A unique key for this model. */ - step?: number; + key: string; /** - * type - * @default collection_slice - * @constant + * Hash + * @description The hash of the model file(s). */ - type: "collection_slice"; - }; - /** - * CollectionSliceOutput - * @description The output of the collection slice node. - */ - CollectionSliceOutput: { + hash: string; /** - * Collection - * @description The sliced collection + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - collection: unknown[]; + path: string; /** - * type - * @default collection_slice_output - * @constant + * File Size + * @description The size of the model in bytes. */ - type: "collection_slice_output"; - }; - /** - * Collection Sort - * @description CollectionSort Sorts a collection - */ - CollectionSortInvocation: { + file_size: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Name + * @description Name of the model. */ - id: string; + name: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Description + * @description Model description */ - is_intermediate?: boolean; + description: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - use_cache?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Collection - * @description collection - * @default [] + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - collection?: unknown[]; + source_api_response: string | null; /** - * Reverse - * @description Reverse Sort - * @default false + * Cover Image + * @description Url for image to preview model */ - reverse?: boolean; + cover_image: string | null; /** - * type - * @default collection_sort + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Type + * @default controlnet * @constant */ - type: "collection_sort"; - }; - /** CollectionSortOutput */ - CollectionSortOutput: { + type: "controlnet"; /** - * Collection - * @description The collection of output items + * Format + * @default checkpoint + * @constant */ - collection: unknown[]; + format: "checkpoint"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * type - * @default collection_sort_output + * Base + * @default sd-1 * @constant */ - type: "collection_sort_output"; + base: "sd-1"; }; - /** - * Collection Unique - * @description Removes duplicate items from a collection. - */ - CollectionUniqueInvocation: { + /** ControlNet_Checkpoint_SD2_Config */ + ControlNet_Checkpoint_SD2_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; - /** - * Collection - * @description The collection to deduplicate - * @default [] - */ - collection?: unknown[]; - /** - * type - * @default collection_unique - * @constant - */ - type: "collection_unique"; - }; - /** - * CollectionUniqueOutput - * @description The output of the collection unique node. - */ - CollectionUniqueOutput: { - /** - * Collection - * @description The collection with unique items - */ - collection: unknown[]; + path: string; /** - * type - * @default collection_unique_output - * @constant + * File Size + * @description The size of the model in bytes. */ - type: "collection_unique_output"; - }; - /** - * Color Cast Correction - * @description Correct color cast while preserving perceptual brightness - */ - ColorCastCorrectionInvocation: { + file_size: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Name + * @description Name of the model. */ - metadata?: components["schemas"]["MetadataField"] | null; + name: string; /** - * @description The board to save the image to - * @default null + * Description + * @description Model description */ - board?: components["schemas"]["BoardField"] | null; + description: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source + * @description The original source of the model (path, URL or repo_id). */ - id: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - is_intermediate?: boolean; + source_api_response: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Cover Image + * @description Url for image to preview model */ - use_cache?: boolean; + cover_image: string | null; /** - * @description Image to correct - * @default null + * Config Path + * @description Path to the config for this model, if any. */ - image?: components["schemas"]["ImageField"] | null; + config_path: string | null; /** - * Strength - * @description Strength of correction adjustment - * @default 0.75 + * Type + * @default controlnet + * @constant */ - strength?: number; + type: "controlnet"; /** - * @description The target color for correction - * @default { - * "r": 127, - * "g": 127, - * "b": 127, - * "a": 255 - * } + * Format + * @default checkpoint + * @constant */ - target_color?: components["schemas"]["ColorField"]; + format: "checkpoint"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * type - * @default color_cast_correction + * Base + * @default sd-2 * @constant */ - type: "color_cast_correction"; + base: "sd-2"; }; - /** - * ColorCollectionOutput - * @description Base class for nodes that output a collection of colors - */ - ColorCollectionOutput: { + /** ControlNet_Checkpoint_SDXL_Config */ + ControlNet_Checkpoint_SDXL_Config: { /** - * Collection - * @description The output colors + * Key + * @description A unique key for this model. */ - collection: components["schemas"]["ColorField"][]; + key: string; /** - * type - * @default color_collection_output - * @constant + * Hash + * @description The hash of the model file(s). */ - type: "color_collection_output"; - }; - /** - * Color Correct - * @description Matches the color histogram of a base image to a reference image, optionally - * using a mask to only color-correct certain regions of the base image. - */ - ColorCorrectInvocation: { + hash: string; /** - * @description The board to save the image to - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - board?: components["schemas"]["BoardField"] | null; + path: string; /** - * @description Optional metadata to be saved with the image - * @default null + * File Size + * @description The size of the model in bytes. */ - metadata?: components["schemas"]["MetadataField"] | null; + file_size: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Name + * @description Name of the model. */ - id: string; + name: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Description + * @description Model description */ - is_intermediate?: boolean; + description: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source + * @description The original source of the model (path, URL or repo_id). */ - use_cache?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * @description The image to color-correct - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - base_image?: components["schemas"]["ImageField"] | null; + source_api_response: string | null; /** - * @description Reference image for color-correction - * @default null + * Cover Image + * @description Url for image to preview model */ - color_reference?: components["schemas"]["ImageField"] | null; + cover_image: string | null; /** - * @description Optional mask to limit color correction area - * @default null + * Config Path + * @description Path to the config for this model, if any. */ - mask?: components["schemas"]["ImageField"] | null; + config_path: string | null; /** - * Color Space - * @description Colorspace in which to apply histogram matching - * @default RGB - * @enum {string} + * Type + * @default controlnet + * @constant */ - colorspace?: "RGB" | "YCbCr" | "YCbCr-Chroma" | "YCbCr-Luma"; + type: "controlnet"; /** - * type - * @default color_correct + * Format + * @default checkpoint * @constant */ - type: "color_correct"; + format: "checkpoint"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; }; /** - * ColorField - * @description A color primitive field + * ControlNet_Checkpoint_ZImage_Config + * @description Model config for Z-Image Control adapter models (Safetensors checkpoint). + * + * Z-Image Control models are standalone adapters containing only the control layers + * (control_layers, control_all_x_embedder, control_noise_refiner) that extend + * the base Z-Image transformer with spatial conditioning capabilities. + * + * Supports: Canny, HED, Depth, Pose, MLSD. + * Recommended control_context_scale: 0.65-0.80. */ - ColorField: { + ControlNet_Checkpoint_ZImage_Config: { /** - * R - * @description The red component + * Key + * @description A unique key for this model. */ - r: number; + key: string; /** - * G - * @description The green component + * Hash + * @description The hash of the model file(s). */ - g: number; + hash: string; /** - * B - * @description The blue component + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - b: number; + path: string; /** - * A - * @description The alpha component + * File Size + * @description The size of the model in bytes. */ - a: number; - }; - /** - * Color Primitive - * @description A color primitive value - */ - ColorInvocation: { + file_size: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Name + * @description Name of the model. */ - id: string; + name: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Description + * @description Model description */ - is_intermediate?: boolean; + description: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source + * @description The original source of the model (path, URL or repo_id). */ - use_cache?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * @description The color value - * @default { - * "r": 0, - * "g": 0, - * "b": 0, - * "a": 255 - * } + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - color?: components["schemas"]["ColorField"]; + source_api_response: string | null; /** - * type - * @default color - * @constant + * Cover Image + * @description Url for image to preview model */ - type: "color"; - }; - /** - * Color Map - * @description Generates a color map from the provided image. - */ - ColorMapInvocation: { + cover_image: string | null; /** - * @description The board to save the image to - * @default null + * Config Path + * @description Path to the config for this model, if any. */ - board?: components["schemas"]["BoardField"] | null; + config_path: string | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Type + * @default controlnet + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "controlnet"; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Format + * @default checkpoint + * @constant */ - id: string; + format: "checkpoint"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Base + * @default z-image + * @constant */ - is_intermediate?: boolean; + base: "z-image"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + }; + /** ControlNet_Diffusers_FLUX_Config */ + ControlNet_Diffusers_FLUX_Config: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Key + * @description A unique key for this model. */ - use_cache?: boolean; + key: string; /** - * @description The image to process - * @default null + * Hash + * @description The hash of the model file(s). */ - image?: components["schemas"]["ImageField"] | null; + hash: string; /** - * Tile Size - * @description Tile size - * @default 64 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - tile_size?: number; + path: string; /** - * type - * @default color_map - * @constant + * File Size + * @description The size of the model in bytes. */ - type: "color_map"; - }; - /** - * ColorOutput - * @description Base class for nodes that output a single color - */ - ColorOutput: { - /** @description The output color */ - color: components["schemas"]["ColorField"]; + file_size: number; /** - * type - * @default color_output - * @constant + * Name + * @description Name of the model. */ - type: "color_output"; - }; - /** - * Compare Floats - * @description Compares two floats based on input criteria and ouputs a boolean. - */ - CompareFloatsInvocation: { + name: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Description + * @description Model description */ - id: string; + description: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - is_intermediate?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - use_cache?: boolean; + source_api_response: string | null; /** - * Comparison Method - * @description The comparision method to use - * @default null + * Cover Image + * @description Url for image to preview model */ - comparison_method?: ("<" | "<=" | ">" | ">=" | "==" | "!=") | null; + cover_image: string | null; /** - * Float 1 - * @description The first float - * @default null + * Format + * @default diffusers + * @constant */ - float1?: number | null; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Float 2 - * @description The second float - * @default null + * Type + * @default controlnet + * @constant */ - float2?: number | null; + type: "controlnet"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * type - * @default compare_floats_invocation + * Base + * @default flux * @constant */ - type: "compare_floats_invocation"; + base: "flux"; }; - /** - * Compare Ints - * @description Compares two integers based on input criteria and ouputs a boolean. - */ - CompareIntsInvocation: { + /** ControlNet_Diffusers_SD1_Config */ + ControlNet_Diffusers_SD1_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Comparison Method - * @description The comparision method to use - * @default null + * File Size + * @description The size of the model in bytes. */ - comparison_method?: ("<" | "<=" | ">" | ">=" | "==" | "!=") | null; + file_size: number; /** - * Int 1 - * @description The first integer. - * @default null + * Name + * @description Name of the model. */ - int1?: number | null; + name: string; /** - * Int 2 - * @description The second integer. - * @default null + * Description + * @description Model description */ - int2?: number | null; + description: string | null; /** - * type - * @default compare_ints_invocation - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "compare_ints_invocation"; - }; - /** - * Compare Strings - * @description Compares two strings based on input criteria and ouputs a boolean. - */ - CompareStringsInvocation: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - id: string; + source_api_response: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Cover Image + * @description Url for image to preview model */ - is_intermediate?: boolean; + cover_image: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Format + * @default diffusers + * @constant */ - use_cache?: boolean; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Comparison Method - * @description The comparision method to use - * @default null + * Type + * @default controlnet + * @constant */ - comparison_method?: ("equals" | "contains" | "starts with" | "ends with") | null; + type: "controlnet"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Ignore Case - * @description If true the node will ignore the case of the strings in all comparison methods - * @default null + * Base + * @default sd-1 + * @constant */ - ignore_case?: boolean | null; + base: "sd-1"; + }; + /** ControlNet_Diffusers_SD2_Config */ + ControlNet_Diffusers_SD2_Config: { /** - * String 1 - * @description The first float - * @default null + * Key + * @description A unique key for this model. */ - str1?: string | null; + key: string; /** - * String 2 - * @description The second float - * @default null + * Hash + * @description The hash of the model file(s). */ - str2?: string | null; + hash: string; /** - * type - * @default compare_strings_invocation - * @constant + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - type: "compare_strings_invocation"; - }; - /** - * Prompt - SD1.5 - * @description Parse prompt using compel package to conditioning. - */ - CompelInvocation: { + path: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * File Size + * @description The size of the model in bytes. */ - id: string; + file_size: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Name + * @description Name of the model. */ - is_intermediate?: boolean; + name: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Description + * @description Model description */ - use_cache?: boolean; + description: string | null; /** - * Prompt - * @description Prompt to be parsed by Compel to create a conditioning tensor - * @default + * Source + * @description The original source of the model (path, URL or repo_id). */ - prompt?: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - clip?: components["schemas"]["CLIPField"] | null; + source_api_response: string | null; /** - * @description A mask defining the region that this conditioning prompt applies to. - * @default null + * Cover Image + * @description Url for image to preview model */ - mask?: components["schemas"]["TensorField"] | null; + cover_image: string | null; /** - * type - * @default compel + * Format + * @default diffusers * @constant */ - type: "compel"; - }; - /** - * Concatenate Flux Conditionings - * @description Concatenates the T5 embedding tensors of up to six input Flux Conditioning objects. - * Provides flexible control over the CLIP embedding: select by 1-indexed input number, - * or generate a zeros tensor. - */ - ConcatenateFluxConditioningInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * @description First optional Flux Conditioning input. - * @default null + * Type + * @default controlnet + * @constant */ - conditioning_1?: components["schemas"]["FluxConditioningField"] | null; + type: "controlnet"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Strength 1 - * @description Strength for the first conditioning input (multiplies its embedding tensors). - * @default 1 + * Base + * @default sd-2 + * @constant */ - strength_1?: number; + base: "sd-2"; + }; + /** ControlNet_Diffusers_SDXL_Config */ + ControlNet_Diffusers_SDXL_Config: { /** - * @description Second optional Flux Conditioning input. - * @default null + * Key + * @description A unique key for this model. */ - conditioning_2?: components["schemas"]["FluxConditioningField"] | null; + key: string; /** - * Strength 2 - * @description Strength for the second conditioning input (multiplies its embedding tensors). - * @default 1 + * Hash + * @description The hash of the model file(s). */ - strength_2?: number; + hash: string; /** - * @description Third optional Flux Conditioning input. - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - conditioning_3?: components["schemas"]["FluxConditioningField"] | null; + path: string; /** - * Strength 3 - * @description Strength for the third conditioning input (multiplies its embedding tensors). - * @default 1 + * File Size + * @description The size of the model in bytes. */ - strength_3?: number; + file_size: number; /** - * @description Fourth optional Flux Conditioning input. - * @default null + * Name + * @description Name of the model. */ - conditioning_4?: components["schemas"]["FluxConditioningField"] | null; + name: string; /** - * Strength 4 - * @description Strength for the fourth conditioning input (multiplies its embedding tensors). - * @default 1 + * Description + * @description Model description */ - strength_4?: number; + description: string | null; /** - * @description Fifth optional Flux Conditioning input. - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - conditioning_5?: components["schemas"]["FluxConditioningField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Strength 5 - * @description Strength for the fifth conditioning input (multiplies its embedding tensors). - * @default 1 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - strength_5?: number; + source_api_response: string | null; /** - * @description Sixth optional Flux Conditioning input. - * @default null + * Cover Image + * @description Url for image to preview model */ - conditioning_6?: components["schemas"]["FluxConditioningField"] | null; + cover_image: string | null; /** - * Strength 6 - * @description Strength for the sixth conditioning input (multiplies its embedding tensors). - * @default 1 + * Format + * @default diffusers + * @constant */ - strength_6?: number; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Select Clip - * @description CLIP embedding selection: 0 for a zeros tensor; 1-6 to select a specific input (1-indexed). If a selected input is missing, it falls back to the next subsequent, then preceding, available CLIP embedding. - * @default 1 + * Type + * @default controlnet + * @constant */ - select_clip?: number; + type: "controlnet"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * type - * @default flux_conditioning_concatenate + * Base + * @default sdxl * @constant */ - type: "flux_conditioning_concatenate"; + base: "sdxl"; }; /** - * Conditioning Collection Primitive - * @description A collection of conditioning tensor primitive values + * ControlOutput + * @description node output for ControlNet info */ - ConditioningCollectionInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The collection of conditioning tensors - * @default [] - */ - collection?: components["schemas"]["ConditioningField"][]; + ControlOutput: { + /** @description ControlNet(s) to apply */ + control: components["schemas"]["ControlField"]; /** * type - * @default conditioning_collection + * @default control_output * @constant */ - type: "conditioning_collection"; + type: "control_output"; }; /** - * Conditioning Collection Primitive Linked - * @description A collection of conditioning tensor primitive values + * Core Metadata + * @description Used internally by Invoke to collect metadata for generations. */ - ConditioningCollectionLinkedInvocation: { + CoreMetadataInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -7518,108 +6467,225 @@ export type components = { */ use_cache?: boolean; /** - * Collection - * @description The collection of conditioning tensors - * @default [] + * Generation Mode + * @description The generation mode that output this image + * @default null */ - collection?: components["schemas"]["ConditioningField"][]; + generation_mode?: ("txt2img" | "img2img" | "inpaint" | "outpaint" | "sdxl_txt2img" | "sdxl_img2img" | "sdxl_inpaint" | "sdxl_outpaint" | "flux_txt2img" | "flux_img2img" | "flux_inpaint" | "flux_outpaint" | "flux2_txt2img" | "flux2_img2img" | "flux2_inpaint" | "flux2_outpaint" | "sd3_txt2img" | "sd3_img2img" | "sd3_inpaint" | "sd3_outpaint" | "cogview4_txt2img" | "cogview4_img2img" | "cogview4_inpaint" | "cogview4_outpaint" | "z_image_txt2img" | "z_image_img2img" | "z_image_inpaint" | "z_image_outpaint") | null; /** - * type - * @default conditioning_collection_linked - * @constant + * Positive Prompt + * @description The positive prompt parameter + * @default null */ - type: "conditioning_collection_linked"; + positive_prompt?: string | null; /** - * @description Conditioning tensor + * Negative Prompt + * @description The negative prompt parameter * @default null */ - conditioning?: components["schemas"]["ConditioningField"] | null; - }; - /** - * ConditioningCollectionOutput - * @description Base class for nodes that output a collection of conditioning tensors - */ - ConditioningCollectionOutput: { + negative_prompt?: string | null; /** - * Collection - * @description The output conditioning tensors + * Width + * @description The width parameter + * @default null */ - collection: components["schemas"]["ConditioningField"][]; + width?: number | null; /** - * type - * @default conditioning_collection_output - * @constant + * Height + * @description The height parameter + * @default null */ - type: "conditioning_collection_output"; - }; - /** - * Conditioning Collection Toggle - * @description Allows boolean selection between two separate conditioning collection inputs - */ - ConditioningCollectionToggleInvocation: { + height?: number | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Seed + * @description The seed used for noise generation + * @default null */ - id: string; + seed?: number | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Rand Device + * @description The device used for random number generation + * @default null */ - is_intermediate?: boolean; + rand_device?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Cfg Scale + * @description The classifier-free guidance scale parameter + * @default null */ - use_cache?: boolean; + cfg_scale?: number | null; /** - * Use Second - * @description Use 2nd Input - * @default false + * Cfg Rescale Multiplier + * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR + * @default null */ - use_second?: boolean; + cfg_rescale_multiplier?: number | null; /** - * Col1 - * @description First Conditioning Collection Input + * Steps + * @description The number of steps used for inference * @default null */ - col1?: components["schemas"]["ConditioningField"][] | null; + steps?: number | null; /** - * Col2 - * @description Second Conditioning Collection Input + * Scheduler + * @description The scheduler used for inference * @default null */ - col2?: components["schemas"]["ConditioningField"][] | null; + scheduler?: string | null; /** - * type - * @default conditioning_collection_toggle - * @constant + * Seamless X + * @description Whether seamless tiling was used on the X axis + * @default null */ - type: "conditioning_collection_toggle"; - }; - /** - * ConditioningField - * @description A conditioning tensor primitive value - */ - ConditioningField: { + seamless_x?: boolean | null; /** - * Conditioning Name - * @description The name of conditioning tensor + * Seamless Y + * @description Whether seamless tiling was used on the Y axis + * @default null */ - conditioning_name: string; + seamless_y?: boolean | null; /** - * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. + * Clip Skip + * @description The number of skipped CLIP layers * @default null */ - mask?: components["schemas"]["TensorField"] | null; + clip_skip?: number | null; + /** + * @description The main model used for inference + * @default null + */ + model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Controlnets + * @description The ControlNets used for inference + * @default null + */ + controlnets?: components["schemas"]["ControlNetMetadataField"][] | null; + /** + * Ipadapters + * @description The IP Adapters used for inference + * @default null + */ + ipAdapters?: components["schemas"]["IPAdapterMetadataField"][] | null; + /** + * T2Iadapters + * @description The IP Adapters used for inference + * @default null + */ + t2iAdapters?: components["schemas"]["T2IAdapterMetadataField"][] | null; + /** + * Loras + * @description The LoRAs used for inference + * @default null + */ + loras?: components["schemas"]["LoRAMetadataField"][] | null; + /** + * Strength + * @description The strength used for latents-to-latents + * @default null + */ + strength?: number | null; + /** + * Init Image + * @description The name of the initial image + * @default null + */ + init_image?: string | null; + /** + * @description The VAE used for decoding, if the main model's default was not used + * @default null + */ + vae?: components["schemas"]["ModelIdentifierField"] | null; + /** + * @description The Qwen3 text encoder model used for Z-Image inference + * @default null + */ + qwen3_encoder?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Hrf Enabled + * @description Whether or not high resolution fix was enabled. + * @default null + */ + hrf_enabled?: boolean | null; + /** + * Hrf Method + * @description The high resolution fix upscale method. + * @default null + */ + hrf_method?: string | null; + /** + * Hrf Strength + * @description The high resolution fix img2img strength used in the upscale pass. + * @default null + */ + hrf_strength?: number | null; + /** + * Positive Style Prompt + * @description The positive style prompt parameter + * @default null + */ + positive_style_prompt?: string | null; + /** + * Negative Style Prompt + * @description The negative style prompt parameter + * @default null + */ + negative_style_prompt?: string | null; + /** + * @description The SDXL Refiner model used + * @default null + */ + refiner_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Refiner Cfg Scale + * @description The classifier-free guidance scale parameter used for the refiner + * @default null + */ + refiner_cfg_scale?: number | null; + /** + * Refiner Steps + * @description The number of steps used for the refiner + * @default null + */ + refiner_steps?: number | null; + /** + * Refiner Scheduler + * @description The scheduler used for the refiner + * @default null + */ + refiner_scheduler?: string | null; + /** + * Refiner Positive Aesthetic Score + * @description The aesthetic score used for the refiner + * @default null + */ + refiner_positive_aesthetic_score?: number | null; + /** + * Refiner Negative Aesthetic Score + * @description The aesthetic score used for the refiner + * @default null + */ + refiner_negative_aesthetic_score?: number | null; + /** + * Refiner Start + * @description The start value used for refiner denoising + * @default null + */ + refiner_start?: number | null; + /** + * type + * @default core_metadata + * @constant + */ + type: "core_metadata"; + } & { + [key: string]: unknown; }; /** - * Conditioning Primitive - * @description A conditioning tensor primitive value + * Create Denoise Mask + * @description Creates mask for denoising model run. */ - ConditioningInvocation: { + CreateDenoiseMaskInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -7638,36 +6704,44 @@ export type components = { */ use_cache?: boolean; /** - * @description Conditioning tensor + * @description VAE * @default null */ - conditioning?: components["schemas"]["ConditioningField"] | null; + vae?: components["schemas"]["VAEField"] | null; /** - * type - * @default conditioning - * @constant + * @description Image which will be masked + * @default null */ - type: "conditioning"; - }; - /** - * ConditioningOutput - * @description Base class for nodes that output a single conditioning tensor - */ - ConditioningOutput: { - /** @description Conditioning tensor */ - conditioning: components["schemas"]["ConditioningField"]; + image?: components["schemas"]["ImageField"] | null; + /** + * @description The mask to use when pasting + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * Tiled + * @description Processing using overlapping tiles (reduce memory consumption) + * @default false + */ + tiled?: boolean; + /** + * Fp32 + * @description Whether or not to use full float32 precision + * @default false + */ + fp32?: boolean; /** * type - * @default conditioning_output + * @default create_denoise_mask * @constant */ - type: "conditioning_output"; + type: "create_denoise_mask"; }; /** - * Conditioning Toggle - * @description Allows boolean selection between two separate conditioning inputs + * Create Gradient Mask + * @description Creates mask for denoising. */ - ConditioningToggleInvocation: { + CreateGradientMaskInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -7686,33 +6760,70 @@ export type components = { */ use_cache?: boolean; /** - * Use Second - * @description Use 2nd Input - * @default false + * @description Image which will be masked + * @default null + */ + mask?: components["schemas"]["ImageField"] | null; + /** + * Edge Radius + * @description How far to expand the edges of the mask + * @default 16 + */ + edge_radius?: number; + /** + * Coherence Mode + * @default Gaussian Blur + * @enum {string} + */ + coherence_mode?: "Gaussian Blur" | "Box Blur" | "Staged"; + /** + * Minimum Denoise + * @description Minimum denoise level for the coherence region + * @default 0 + */ + minimum_denoise?: number; + /** + * [OPTIONAL] Image + * @description OPTIONAL: Only connect for specialized Inpainting models, masked_latents will be generated from the image with the VAE + * @default null */ - use_second?: boolean; + image?: components["schemas"]["ImageField"] | null; /** - * @description First Conditioning Input + * [OPTIONAL] UNet + * @description OPTIONAL: If the Unet is a specialized Inpainting model, masked_latents will be generated from the image with the VAE * @default null */ - cond1?: components["schemas"]["ConditioningField"] | null; + unet?: components["schemas"]["UNetField"] | null; /** - * @description Second Conditioning Input + * [OPTIONAL] VAE + * @description OPTIONAL: Only connect for specialized Inpainting models, masked_latents will be generated from the image with the VAE * @default null */ - cond2?: components["schemas"]["ConditioningField"] | null; + vae?: components["schemas"]["VAEField"] | null; + /** + * Tiled + * @description Processing using overlapping tiles (reduce memory consumption) + * @default false + */ + tiled?: boolean; + /** + * Fp32 + * @description Whether or not to use full float32 precision + * @default false + */ + fp32?: boolean; /** * type - * @default conditioning_toggle + * @default create_gradient_mask * @constant */ - type: "conditioning_toggle"; + type: "create_gradient_mask"; }; /** - * Content Shuffle - * @description Shuffles the image, similar to a 'liquify' filter. + * Crop Image to Bounding Box + * @description Crop an image to the given bounding box. If the bounding box is omitted, the image is cropped to the non-transparent pixels. */ - ContentShuffleInvocation: { + CropImageToBoundingBoxInvocation: { /** * @description The board to save the image to * @default null @@ -7741,172 +6852,145 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description The image to crop * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Scale Factor - * @description The scale factor used for the shuffle - * @default 256 + * @description The bounding box to crop the image to + * @default null */ - scale_factor?: number; + bounding_box?: components["schemas"]["BoundingBoxField"] | null; /** * type - * @default content_shuffle + * @default crop_image_to_bounding_box * @constant */ - type: "content_shuffle"; - }; - /** ControlAdapterDefaultSettings */ - ControlAdapterDefaultSettings: { - /** Preprocessor */ - preprocessor: string | null; + type: "crop_image_to_bounding_box"; }; - /** ControlField */ - ControlField: { - /** @description The control image */ - image: components["schemas"]["ImageField"]; - /** @description The ControlNet model to use */ - control_model: components["schemas"]["ModelIdentifierField"]; + /** + * Crop Latents + * @description Crops a latent-space tensor to a box specified in image-space. The box dimensions and coordinates must be + * divisible by the latent scale factor of 8. + */ + CropLatentsCoreInvocation: { /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - control_weight?: number | number[]; + id: string; /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - begin_step_percent?: number; + is_intermediate?: boolean; /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - end_step_percent?: number; + use_cache?: boolean; /** - * Control Mode - * @description The control mode to use - * @default balanced - * @enum {string} + * @description Latents tensor + * @default null */ - control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + latents?: components["schemas"]["LatentsField"] | null; /** - * Resize Mode - * @description The resize mode to use - * @default just_resize - * @enum {string} + * X + * @description The left x coordinate (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @default null */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; - }; - /** ControlListOutput */ - ControlListOutput: { + x?: number | null; /** - * ControlNet List - * @description ControlNet(s) to apply + * Y + * @description The top y coordinate (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @default null + */ + y?: number | null; + /** + * Width + * @description The width (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @default null */ - control_list: components["schemas"]["ControlField"][]; + width?: number | null; /** - * type - * @default control_list_output - * @constant + * Height + * @description The height (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. + * @default null */ - type: "control_list_output"; - }; - /** ControlLoRAField */ - ControlLoRAField: { - /** @description Info to load lora model */ - lora: components["schemas"]["ModelIdentifierField"]; + height?: number | null; /** - * Weight - * @description Weight to apply to lora model + * type + * @default crop_latents + * @constant */ - weight: number; - /** @description Image to use in structural conditioning */ - img: components["schemas"]["ImageField"]; + type: "crop_latents"; }; /** - * ControlLoRA_LyCORIS_FLUX_Config - * @description Model config for Control LoRA models. + * OpenCV Inpaint + * @description Simple inpaint using opencv. */ - ControlLoRA_LyCORIS_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; + CvInpaintInvocation: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * @description The board to save the image to + * @default null */ - path: string; + board?: components["schemas"]["BoardField"] | null; /** - * File Size - * @description The size of the model in bytes. + * @description Optional metadata to be saved with the image + * @default null */ - file_size: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Name - * @description Name of the model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - name: string; + id: string; /** - * Description - * @description Model description + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - description: string | null; + is_intermediate?: boolean; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + use_cache?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * @description The image to inpaint + * @default null */ - source_api_response: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Cover Image - * @description Url for image to preview model + * @description The mask to use when inpainting + * @default null */ - cover_image: string | null; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + mask?: components["schemas"]["ImageField"] | null; /** - * Base - * @default flux + * type + * @default cv_inpaint * @constant */ - base: "flux"; + type: "cv_inpaint"; + }; + /** + * DW Openpose Detection + * @description Generates an openpose pose from an image using DWPose + */ + DWOpenposeDetectionInvocation: { /** - * Type - * @default control_lora - * @constant + * @description The board to save the image to + * @default null */ - type: "control_lora"; + board?: components["schemas"]["BoardField"] | null; /** - * Format - * @default lycoris - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - format: "lycoris"; - /** Trigger Phrases */ - trigger_phrases: string[] | null; - }; - /** - * ControlNet - SD1.5, SD2, SDXL - * @description Collects ControlNet info to pass to other nodes - */ - ControlNetInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -7925,59 +7009,37 @@ export type components = { */ use_cache?: boolean; /** - * @description The control image + * @description The image to process * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description ControlNet model to load - * @default null - */ - control_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 - */ - control_weight?: number | number[]; - /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 - */ - begin_step_percent?: number; - /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Draw Body + * @default true */ - end_step_percent?: number; + draw_body?: boolean; /** - * Control Mode - * @description The control mode used - * @default balanced - * @enum {string} + * Draw Face + * @default false */ - control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + draw_face?: boolean; /** - * Resize Mode - * @description The resize mode used - * @default just_resize - * @enum {string} + * Draw Hands + * @default false */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + draw_hands?: boolean; /** * type - * @default controlnet + * @default dw_openpose_detection * @constant */ - type: "controlnet"; + type: "dw_openpose_detection"; }; /** - * ControlNet-Linked - * @description Collects ControlNet info to pass to other nodes. + * Decode Invisible Watermark + * @description Decode an invisible watermark from an image. */ - ControlNetLinkedInvocation: { + DecodeInvisibleWatermarkInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -7996,908 +7058,913 @@ export type components = { */ use_cache?: boolean; /** - * @description The control image + * @description The image to decode the watermark from * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description ControlNet model to load - * @default null + * Length + * @description The expected watermark length in bytes + * @default 8 */ - control_model?: components["schemas"]["ModelIdentifierField"] | null; + length?: number; /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * type + * @default decode_watermark + * @constant */ - control_weight?: number | number[]; - /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 - */ - begin_step_percent?: number; - /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 - */ - end_step_percent?: number; + type: "decode_watermark"; + }; + /** + * DeleteAllExceptCurrentResult + * @description Result of deleting all except current + */ + DeleteAllExceptCurrentResult: { /** - * Control Mode - * @description The control mode used - * @default balanced - * @enum {string} + * Deleted + * @description Number of queue items deleted */ - control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + deleted: number; + }; + /** DeleteBoardResult */ + DeleteBoardResult: { /** - * Resize Mode - * @description The resize mode used - * @default just_resize - * @enum {string} + * Board Id + * @description The id of the board that was deleted. */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + board_id: string; /** - * type - * @default controlnet-linked - * @constant + * Deleted Board Images + * @description The image names of the board-images relationships that were deleted. */ - type: "controlnet-linked"; + deleted_board_images: string[]; /** - * ControlNet-List - * @description ControlNet(s) to apply - * @default null + * Deleted Images + * @description The names of the images that were deleted. */ - control_list?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + deleted_images: string[]; }; - /** ControlNetMetadataField */ - ControlNetMetadataField: { - /** @description The control image */ - image: components["schemas"]["ImageField"]; + /** + * DeleteByDestinationResult + * @description Result of deleting by a destination + */ + DeleteByDestinationResult: { /** - * @description The control image, after processing. - * @default null + * Deleted + * @description Number of queue items deleted */ - processed_image?: components["schemas"]["ImageField"] | null; - /** @description The ControlNet model to use */ - control_model: components["schemas"]["ModelIdentifierField"]; + deleted: number; + }; + /** DeleteImagesResult */ + DeleteImagesResult: { /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * Affected Boards + * @description The ids of boards affected by the delete operation */ - control_weight?: number | number[]; + affected_boards: string[]; /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 + * Deleted Images + * @description The names of the images that were deleted */ - begin_step_percent?: number; + deleted_images: string[]; + }; + /** + * DeleteOrphanedModelsRequest + * @description Request to delete specific orphaned model directories. + */ + DeleteOrphanedModelsRequest: { /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Paths + * @description List of relative paths to delete */ - end_step_percent?: number; + paths: string[]; + }; + /** + * DeleteOrphanedModelsResponse + * @description Response from deleting orphaned models. + */ + DeleteOrphanedModelsResponse: { /** - * Control Mode - * @description The control mode to use - * @default balanced - * @enum {string} + * Deleted + * @description Paths that were successfully deleted */ - control_mode?: "balanced" | "more_prompt" | "more_control" | "unbalanced"; + deleted: string[]; /** - * Resize Mode - * @description The resize mode to use - * @default just_resize - * @enum {string} + * Errors + * @description Paths that had errors, with error messages */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + errors: { + [key: string]: string; + }; }; /** - * ControlNetRecallParameter - * @description ControlNet configuration for recall + * Denoise - SD1.5, SDXL + * @description Denoises noisy latents to decodable images */ - ControlNetRecallParameter: { + DenoiseLatentsInvocation: { /** - * Model Name - * @description The name of the ControlNet/T2I Adapter/Control LoRA model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - model_name: string; + id: string; /** - * Image Name - * @description The filename of the control image in outputs/images + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - image_name?: string | null; + is_intermediate?: boolean; /** - * Weight - * @description The weight for the control adapter - * @default 1 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - weight?: number; + use_cache?: boolean; /** - * Begin Step Percent - * @description When the control adapter is first applied (% of total steps) + * Positive Conditioning + * @description Positive conditioning tensor + * @default null */ - begin_step_percent?: number | null; + positive_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; /** - * End Step Percent - * @description When the control adapter is last applied (% of total steps) + * Negative Conditioning + * @description Negative conditioning tensor + * @default null */ - end_step_percent?: number | null; + negative_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; /** - * Control Mode - * @description The control mode (ControlNet only) + * @description Noise tensor + * @default null */ - control_mode?: ("balanced" | "more_prompt" | "more_control") | null; - }; - /** ControlNet_Checkpoint_FLUX_Config */ - ControlNet_Checkpoint_FLUX_Config: { + noise?: components["schemas"]["LatentsField"] | null; /** - * Key - * @description A unique key for this model. + * Steps + * @description Number of steps to run + * @default 10 */ - key: string; + steps?: number; /** - * Hash - * @description The hash of the model file(s). + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 7.5 */ - hash: string; + cfg_scale?: number | number[]; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - path: string; + denoising_start?: number; /** - * File Size - * @description The size of the model in bytes. + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - file_size: number; + denoising_end?: number; /** - * Name - * @description Name of the model. + * Scheduler + * @description Scheduler to use during inference + * @default euler + * @enum {string} */ - name: string; + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** - * Description - * @description Model description + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - description: string | null; + unet?: components["schemas"]["UNetField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Control + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * IP-Adapter + * @description IP-Adapter to apply + * @default null */ - source_api_response: string | null; + ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * Cover Image - * @description Url for image to preview model + * T2I-Adapter + * @description T2I-Adapter(s) to apply + * @default null */ - cover_image: string | null; + t2i_adapter?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; /** - * Config Path - * @description Path to the config for this model, if any. + * CFG Rescale Multiplier + * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR + * @default 0 */ - config_path: string | null; + cfg_rescale_multiplier?: number; /** - * Type - * @default controlnet - * @constant + * @description Latents tensor + * @default null */ - type: "controlnet"; + latents?: components["schemas"]["LatentsField"] | null; /** - * Format - * @default checkpoint - * @constant + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - format: "checkpoint"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Base - * @default flux + * type + * @default denoise_latents * @constant */ - base: "flux"; + type: "denoise_latents"; }; - /** ControlNet_Checkpoint_SD1_Config */ - ControlNet_Checkpoint_SD1_Config: { + /** Denoise - SD1.5, SDXL + Metadata */ + DenoiseLatentsMetaInvocation: { /** - * Key - * @description A unique key for this model. + * @description Optional metadata to be saved with the image + * @default null */ - key: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Hash - * @description The hash of the model file(s). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - hash: string; + id: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - path: string; + is_intermediate?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - file_size: number; + use_cache?: boolean; /** - * Name - * @description Name of the model. + * Positive Conditioning + * @description Positive conditioning tensor + * @default null */ - name: string; + positive_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; /** - * Description - * @description Model description + * Negative Conditioning + * @description Negative conditioning tensor + * @default null */ - description: string | null; + negative_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * @description Noise tensor + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + noise?: components["schemas"]["LatentsField"] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Steps + * @description Number of steps to run + * @default 10 */ - source_api_response: string | null; + steps?: number; /** - * Cover Image - * @description Url for image to preview model + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 7.5 */ - cover_image: string | null; + cfg_scale?: number | number[]; /** - * Config Path - * @description Path to the config for this model, if any. + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - config_path: string | null; + denoising_start?: number; /** - * Type - * @default controlnet - * @constant + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - type: "controlnet"; + denoising_end?: number; /** - * Format - * @default checkpoint - * @constant + * Scheduler + * @description Scheduler to use during inference + * @default euler + * @enum {string} */ - format: "checkpoint"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** - * Base - * @default sd-1 - * @constant + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - base: "sd-1"; - }; - /** ControlNet_Checkpoint_SD2_Config */ - ControlNet_Checkpoint_SD2_Config: { + unet?: components["schemas"]["UNetField"] | null; /** - * Key - * @description A unique key for this model. + * Control + * @default null */ - key: string; + control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; /** - * Hash - * @description The hash of the model file(s). + * IP-Adapter + * @description IP-Adapter to apply + * @default null */ - hash: string; + ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * T2I-Adapter + * @description T2I-Adapter(s) to apply + * @default null */ - path: string; + t2i_adapter?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; /** - * File Size - * @description The size of the model in bytes. + * CFG Rescale Multiplier + * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR + * @default 0 */ - file_size: number; + cfg_rescale_multiplier?: number; /** - * Name - * @description Name of the model. + * @description Latents tensor + * @default null */ - name: string; + latents?: components["schemas"]["LatentsField"] | null; /** - * Description - * @description Model description + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - description: string | null; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default denoise_latents_meta + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "denoise_latents_meta"; + }; + /** + * DenoiseMaskField + * @description An inpaint mask field + */ + DenoiseMaskField: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Mask Name + * @description The name of the mask image */ - source_api_response: string | null; + mask_name: string; /** - * Cover Image - * @description Url for image to preview model + * Masked Latents Name + * @description The name of the masked image latents + * @default null */ - cover_image: string | null; + masked_latents_name?: string | null; /** - * Config Path - * @description Path to the config for this model, if any. + * Gradient + * @description Used for gradient inpainting + * @default false */ - config_path: string | null; + gradient?: boolean; + }; + /** + * DenoiseMaskOutput + * @description Base class for nodes that output a single image + */ + DenoiseMaskOutput: { + /** @description Mask for denoise model run */ + denoise_mask: components["schemas"]["DenoiseMaskField"]; /** - * Type - * @default controlnet + * type + * @default denoise_mask_output * @constant */ - type: "controlnet"; + type: "denoise_mask_output"; + }; + /** + * Depth Anything Depth Estimation + * @description Generates a depth map using a Depth Anything model. + */ + DepthAnythingDepthEstimationInvocation: { /** - * Format - * @default checkpoint - * @constant + * @description The board to save the image to + * @default null */ - format: "checkpoint"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + board?: components["schemas"]["BoardField"] | null; /** - * Base - * @default sd-2 - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - base: "sd-2"; - }; - /** ControlNet_Checkpoint_SDXL_Config */ - ControlNet_Checkpoint_SDXL_Config: { + metadata?: components["schemas"]["MetadataField"] | null; /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * @description The image to process + * @default null */ - file_size: number; + image?: components["schemas"]["ImageField"] | null; /** - * Name - * @description Name of the model. + * Model Size + * @description The size of the depth model to use + * @default small_v2 + * @enum {string} */ - name: string; + model_size?: "large" | "base" | "small" | "small_v2"; /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default depth_anything_depth_estimation + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "depth_anything_depth_estimation"; + }; + /** + * Divide Integers + * @description Divides two numbers + */ + DivideInvocation: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source_api_response: string | null; + id: string; /** - * Cover Image - * @description Url for image to preview model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cover_image: string | null; + is_intermediate?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - config_path: string | null; + use_cache?: boolean; /** - * Type - * @default controlnet - * @constant + * A + * @description The first number + * @default 0 */ - type: "controlnet"; + a?: number; /** - * Format - * @default checkpoint - * @constant + * B + * @description The second number + * @default 0 */ - format: "checkpoint"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + b?: number; /** - * Base - * @default sdxl + * type + * @default div * @constant */ - base: "sdxl"; + type: "div"; }; /** - * ControlNet_Checkpoint_ZImage_Config - * @description Model config for Z-Image Control adapter models (Safetensors checkpoint). - * - * Z-Image Control models are standalone adapters containing only the control layers - * (control_layers, control_all_x_embedder, control_noise_refiner) that extend - * the base Z-Image transformer with spatial conditioning capabilities. - * - * Supports: Canny, HED, Depth, Pose, MLSD. - * Recommended control_context_scale: 0.65-0.80. + * DownloadCancelledEvent + * @description Event model for download_cancelled */ - ControlNet_Checkpoint_ZImage_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; + DownloadCancelledEvent: { /** - * Hash - * @description The hash of the model file(s). + * Timestamp + * @description The timestamp of the event */ - hash: string; + timestamp: number; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Source + * @description The source of the download */ - path: string; + source: string; + }; + /** + * DownloadCompleteEvent + * @description Event model for download_complete + */ + DownloadCompleteEvent: { /** - * File Size - * @description The size of the model in bytes. + * Timestamp + * @description The timestamp of the event */ - file_size: number; + timestamp: number; /** - * Name - * @description Name of the model. + * Source + * @description The source of the download */ - name: string; + source: string; /** - * Description - * @description Model description + * Download Path + * @description The local path where the download is saved */ - description: string | null; + download_path: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Total Bytes + * @description The total number of bytes downloaded */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + total_bytes: number; + }; + /** + * DownloadErrorEvent + * @description Event model for download_error + */ + DownloadErrorEvent: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Timestamp + * @description The timestamp of the event */ - source_api_response: string | null; + timestamp: number; /** - * Cover Image - * @description Url for image to preview model + * Source + * @description The source of the download */ - cover_image: string | null; + source: string; /** - * Config Path - * @description Path to the config for this model, if any. + * Error Type + * @description The type of error */ - config_path: string | null; + error_type: string; /** - * Type - * @default controlnet - * @constant + * Error + * @description The error message */ - type: "controlnet"; + error: string; + }; + /** + * DownloadJob + * @description Class to monitor and control a model download request. + */ + DownloadJob: { /** - * Format - * @default checkpoint - * @constant + * Id + * @description Numeric ID of this job + * @default -1 */ - format: "checkpoint"; + id?: number; /** - * Base - * @default z-image - * @constant + * Dest + * Format: path + * @description Initial destination of downloaded model on local disk; a directory or file path */ - base: "z-image"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; - }; - /** ControlNet_Diffusers_FLUX_Config */ - ControlNet_Diffusers_FLUX_Config: { + dest: string; /** - * Key - * @description A unique key for this model. + * Download Path + * @description Final location of downloaded file or directory */ - key: string; + download_path?: string | null; /** - * Hash - * @description The hash of the model file(s). + * @description Status of the download + * @default waiting */ - hash: string; + status?: components["schemas"]["DownloadJobStatus"]; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Bytes + * @description Bytes downloaded so far + * @default 0 */ - path: string; + bytes?: number; /** - * File Size - * @description The size of the model in bytes. + * Total Bytes + * @description Total file size (bytes) + * @default 0 */ - file_size: number; + total_bytes?: number; /** - * Name - * @description Name of the model. + * Error Type + * @description Name of exception that caused an error */ - name: string; + error_type?: string | null; /** - * Description - * @description Model description + * Error + * @description Traceback of the exception that caused an error */ - description: string | null; + error?: string | null; /** * Source - * @description The original source of the model (path, URL or repo_id). + * Format: uri + * @description Where to download from. Specific types specified in child classes. */ source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Access Token + * @description authorization token for protected resources */ - source_api_response: string | null; + access_token?: string | null; /** - * Cover Image - * @description Url for image to preview model + * Priority + * @description Queue priority; lower values are higher priority + * @default 10 */ - cover_image: string | null; + priority?: number; /** - * Format - * @default diffusers - * @constant + * Job Started + * @description Timestamp for when the download job started */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + job_started?: string | null; /** - * Type - * @default controlnet - * @constant + * Job Ended + * @description Timestamp for when the download job ende1d (completed or errored) */ - type: "controlnet"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + job_ended?: string | null; /** - * Base - * @default flux - * @constant + * Content Type + * @description Content type of downloaded file */ - base: "flux"; - }; - /** ControlNet_Diffusers_SD1_Config */ - ControlNet_Diffusers_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; + content_type?: string | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Canonical Url + * @description Canonical URL to request on resume */ - path: string; + canonical_url?: string | null; /** - * File Size - * @description The size of the model in bytes. + * Etag + * @description ETag from the remote server, if available */ - file_size: number; + etag?: string | null; /** - * Name - * @description Name of the model. + * Last Modified + * @description Last-Modified from the remote server, if available */ - name: string; + last_modified?: string | null; /** - * Description - * @description Model description + * Final Url + * @description Final resolved URL after redirects, if available */ - description: string | null; + final_url?: string | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Expected Total Bytes + * @description Expected total size of the download */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + expected_total_bytes?: number | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Resume Required + * @description True if server refused resume; restart required + * @default false */ - source_api_response: string | null; + resume_required?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Resume Message + * @description Message explaining why resume is required */ - cover_image: string | null; + resume_message?: string | null; /** - * Format - * @default diffusers - * @constant + * Resume From Scratch + * @description True if resume metadata existed but the partial file was missing and the download restarted from the beginning + * @default false */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + resume_from_scratch?: boolean; + }; + /** + * DownloadJobStatus + * @description State of a download job. + * @enum {string} + */ + DownloadJobStatus: "waiting" | "running" | "paused" | "completed" | "cancelled" | "error"; + /** + * DownloadPausedEvent + * @description Event model for download_paused + */ + DownloadPausedEvent: { /** - * Type - * @default controlnet - * @constant + * Timestamp + * @description The timestamp of the event */ - type: "controlnet"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + timestamp: number; /** - * Base - * @default sd-1 - * @constant + * Source + * @description The source of the download */ - base: "sd-1"; + source: string; }; - /** ControlNet_Diffusers_SD2_Config */ - ControlNet_Diffusers_SD2_Config: { + /** + * DownloadProgressEvent + * @description Event model for download_progress + */ + DownloadProgressEvent: { /** - * Key - * @description A unique key for this model. + * Timestamp + * @description The timestamp of the event */ - key: string; + timestamp: number; /** - * Hash - * @description The hash of the model file(s). + * Source + * @description The source of the download */ - hash: string; + source: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Download Path + * @description The local path where the download is saved */ - path: string; + download_path: string; /** - * File Size - * @description The size of the model in bytes. + * Current Bytes + * @description The number of bytes downloaded so far */ - file_size: number; + current_bytes: number; /** - * Name - * @description Name of the model. + * Total Bytes + * @description The total number of bytes to be downloaded */ - name: string; + total_bytes: number; + }; + /** + * DownloadStartedEvent + * @description Event model for download_started + */ + DownloadStartedEvent: { /** - * Description - * @description Model description + * Timestamp + * @description The timestamp of the event */ - description: string | null; + timestamp: number; /** * Source - * @description The original source of the model (path, URL or repo_id). + * @description The source of the download */ source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Download Path + * @description The local path where the download is saved */ - source_api_response: string | null; + download_path: string; + }; + /** + * Dynamic Prompt + * @description Parses a prompt using adieyal/dynamicprompts' random or combinatorial generator + */ + DynamicPromptInvocation: { /** - * Cover Image - * @description Url for image to preview model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - cover_image: string | null; + id: string; /** - * Format - * @default diffusers - * @constant + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + is_intermediate?: boolean; /** - * Type - * @default controlnet - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default false */ - type: "controlnet"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + use_cache?: boolean; /** - * Base - * @default sd-2 - * @constant + * Prompt + * @description The prompt to parse with dynamicprompts + * @default null */ - base: "sd-2"; - }; - /** ControlNet_Diffusers_SDXL_Config */ - ControlNet_Diffusers_SDXL_Config: { + prompt?: string | null; /** - * Key - * @description A unique key for this model. + * Max Prompts + * @description The number of prompts to generate + * @default 1 */ - key: string; + max_prompts?: number; /** - * Hash - * @description The hash of the model file(s). + * Combinatorial + * @description Whether to use the combinatorial generator + * @default false */ - hash: string; + combinatorial?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * type + * @default dynamic_prompt + * @constant */ - path: string; + type: "dynamic_prompt"; + }; + /** DynamicPromptsResponse */ + DynamicPromptsResponse: { + /** Prompts */ + prompts: string[]; + /** Error */ + error?: string | null; + }; + /** + * Upscale (RealESRGAN) + * @description Upscales an image using RealESRGAN. + */ + ESRGANInvocation: { /** - * File Size - * @description The size of the model in bytes. + * @description The board to save the image to + * @default null */ - file_size: number; + board?: components["schemas"]["BoardField"] | null; /** - * Name - * @description Name of the model. + * @description Optional metadata to be saved with the image + * @default null */ - name: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Format - * @default diffusers - * @constant + * @description The input image + * @default null */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + image?: components["schemas"]["ImageField"] | null; /** - * Type - * @default controlnet - * @constant + * Model Name + * @description The Real-ESRGAN model to use + * @default RealESRGAN_x4plus.pth + * @enum {string} */ - type: "controlnet"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + model_name?: "RealESRGAN_x4plus.pth" | "RealESRGAN_x4plus_anime_6B.pth" | "ESRGAN_SRx4_DF2KOST_official-ff704c30.pth" | "RealESRGAN_x2plus.pth"; /** - * Base - * @default sdxl - * @constant + * Tile Size + * @description Tile size for tiled ESRGAN upscaling (0=tiling disabled) + * @default 400 */ - base: "sdxl"; - }; - /** - * ControlOutput - * @description node output for ControlNet info - */ - ControlOutput: { - /** @description ControlNet(s) to apply */ - control: components["schemas"]["ControlField"]; + tile_size?: number; /** * type - * @default control_output + * @default esrgan * @constant */ - type: "control_output"; + type: "esrgan"; }; - /** - * Coordinated Noise (Flux) - * @description Generates latent noise that is stable for the given coordinates. - * - * That is, the noise at channel=1 x=3 y=4 for seed=42 will always be the same, regardless of the - * total size (width and height) of the region. - * - * It makes reproducible noise that you're able to crop or scroll. - * - * This variant has 16 channels for Flux. - */ - CoordinatedFluxNoiseInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + /** Edge */ + Edge: { + /** @description The connection for the edge's from node and field */ + source: components["schemas"]["EdgeConnection"]; + /** @description The connection for the edge's to node and field */ + destination: components["schemas"]["EdgeConnection"]; + }; + /** EdgeConnection */ + EdgeConnection: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Node Id + * @description The id of the node for this edge connection */ - use_cache?: boolean; + node_id: string; /** - * Seed - * @description Seed for random number generation - * @default null + * Field + * @description The field for this connection */ - seed?: number | null; + field: string; + }; + /** EnqueueBatchResult */ + EnqueueBatchResult: { /** - * Width - * @description Width of output (px) - * @default 512 + * Queue Id + * @description The ID of the queue */ - width?: number; + queue_id: string; /** - * Height - * @description Height of output (px) - * @default 512 + * Enqueued + * @description The total number of queue items enqueued */ - height?: number; + enqueued: number; /** - * X Offset - * @description x-coordinate of the lower edge - * @default 0 + * Requested + * @description The total number of queue items requested to be enqueued */ - x_offset?: number; + requested: number; + /** @description The batch that was enqueued */ + batch: components["schemas"]["Batch"]; /** - * Y Offset - * @description y-coordinate of the lower edge - * @default 0 + * Priority + * @description The priority of the enqueued batch */ - y_offset?: number; + priority: number; /** - * Channel Offset - * @description coordinate of the first channel - * @default 0 + * Item Ids + * @description The IDs of the queue items that were enqueued */ - channel_offset?: number; + item_ids: number[]; + }; + /** + * Expand Mask with Fade + * @description Expands a mask with a fade effect. The mask uses black to indicate areas to keep from the generated image and white for areas to discard. + * The mask is thresholded to create a binary mask, and then a distance transform is applied to create a fade effect. + * The fade size is specified in pixels, and the mask is expanded by that amount. The result is a mask with a smooth transition from black to white. + * If the fade size is 0, the mask is returned as-is. + */ + ExpandMaskWithFadeInvocation: { /** - * Algorithm - * @description psuedo-random generation algorithm - * @default pcg64 - * @enum {string} + * @description The board to save the image to + * @default null */ - algorithm?: "pcg64" | "morton+farmhash"; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default noise_coordinated_flux - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "noise_coordinated_flux"; - }; - /** - * Coordinated Noise - * @description Generates latent noise that is stable for the given coordinates. - * - * That is, the noise at channel=1 x=3 y=4 for seed=42 will always be the same, regardless of the - * total size (width and height) of the region. - * - * It makes reproducible noise that you're able to crop or scroll. - */ - CoordinatedNoiseInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8916,60 +7983,41 @@ export type components = { */ use_cache?: boolean; /** - * Seed - * @description Seed for random number generation + * @description The mask to expand * @default null */ - seed?: number | null; - /** - * Width - * @description Width of output (px) - * @default 512 - */ - width?: number; - /** - * Height - * @description Height of output (px) - * @default 512 - */ - height?: number; - /** - * X Offset - * @description x-coordinate of the lower edge - * @default 0 - */ - x_offset?: number; - /** - * Y Offset - * @description y-coordinate of the lower edge - * @default 0 - */ - y_offset?: number; + mask?: components["schemas"]["ImageField"] | null; /** - * Channel Offset - * @description coordinate of the first channel + * Threshold + * @description The threshold for the binary mask (0-255) * @default 0 */ - channel_offset?: number; + threshold?: number; /** - * Algorithm - * @description psuedo-random generation algorithm - * @default pcg64 - * @enum {string} + * Fade Size Px + * @description The size of the fade in pixels + * @default 32 */ - algorithm?: "pcg64" | "morton+farmhash"; + fade_size_px?: number; /** * type - * @default noise_coordinated + * @default expand_mask_with_fade * @constant */ - type: "noise_coordinated"; + type: "expand_mask_with_fade"; + }; + /** ExposedField */ + ExposedField: { + /** Nodeid */ + nodeId: string; + /** Fieldname */ + fieldName: string; }; /** - * Core Metadata - * @description Used internally by Invoke to collect metadata for generations. + * Apply LoRA Collection - FLUX + * @description Applies a collection of LoRAs to a FLUX transformer. */ - CoreMetadataInvocation: { + FLUXLoRACollectionLoader: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -8988,225 +8036,122 @@ export type components = { */ use_cache?: boolean; /** - * Generation Mode - * @description The generation mode that output this image + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. * @default null */ - generation_mode?: ("txt2img" | "img2img" | "inpaint" | "outpaint" | "sdxl_txt2img" | "sdxl_img2img" | "sdxl_inpaint" | "sdxl_outpaint" | "flux_txt2img" | "flux_img2img" | "flux_inpaint" | "flux_outpaint" | "flux2_txt2img" | "flux2_img2img" | "flux2_inpaint" | "flux2_outpaint" | "sd3_txt2img" | "sd3_img2img" | "sd3_inpaint" | "sd3_outpaint" | "cogview4_txt2img" | "cogview4_img2img" | "cogview4_inpaint" | "cogview4_outpaint" | "z_image_txt2img" | "z_image_img2img" | "z_image_inpaint" | "z_image_outpaint") | null; + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * Positive Prompt - * @description The positive prompt parameter + * Transformer + * @description Transformer * @default null */ - positive_prompt?: string | null; + transformer?: components["schemas"]["TransformerField"] | null; /** - * Negative Prompt - * @description The negative prompt parameter + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - negative_prompt?: string | null; + clip?: components["schemas"]["CLIPField"] | null; /** - * Width - * @description The width parameter + * T5 Encoder + * @description T5 tokenizer and text encoder * @default null */ - width?: number | null; + t5_encoder?: components["schemas"]["T5EncoderField"] | null; /** - * Height - * @description The height parameter - * @default null + * type + * @default flux_lora_collection_loader + * @constant */ - height?: number | null; + type: "flux_lora_collection_loader"; + }; + /** + * FLUXRedux_Checkpoint_Config + * @description Model config for FLUX Tools Redux model. + */ + FLUXRedux_Checkpoint_Config: { /** - * Seed - * @description The seed used for noise generation - * @default null + * Key + * @description A unique key for this model. */ - seed?: number | null; + key: string; /** - * Rand Device - * @description The device used for random number generation - * @default null + * Hash + * @description The hash of the model file(s). */ - rand_device?: string | null; + hash: string; /** - * Cfg Scale - * @description The classifier-free guidance scale parameter - * @default null - */ - cfg_scale?: number | null; - /** - * Cfg Rescale Multiplier - * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR - * @default null - */ - cfg_rescale_multiplier?: number | null; - /** - * Steps - * @description The number of steps used for inference - * @default null - */ - steps?: number | null; - /** - * Scheduler - * @description The scheduler used for inference - * @default null - */ - scheduler?: string | null; - /** - * Seamless X - * @description Whether seamless tiling was used on the X axis - * @default null - */ - seamless_x?: boolean | null; - /** - * Seamless Y - * @description Whether seamless tiling was used on the Y axis - * @default null - */ - seamless_y?: boolean | null; - /** - * Clip Skip - * @description The number of skipped CLIP layers - * @default null - */ - clip_skip?: number | null; - /** - * @description The main model used for inference - * @default null - */ - model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Controlnets - * @description The ControlNets used for inference - * @default null - */ - controlnets?: components["schemas"]["ControlNetMetadataField"][] | null; - /** - * Ipadapters - * @description The IP Adapters used for inference - * @default null - */ - ipAdapters?: components["schemas"]["IPAdapterMetadataField"][] | null; - /** - * T2Iadapters - * @description The IP Adapters used for inference - * @default null - */ - t2iAdapters?: components["schemas"]["T2IAdapterMetadataField"][] | null; - /** - * Loras - * @description The LoRAs used for inference - * @default null - */ - loras?: components["schemas"]["LoRAMetadataField"][] | null; - /** - * Strength - * @description The strength used for latents-to-latents - * @default null - */ - strength?: number | null; - /** - * Init Image - * @description The name of the initial image - * @default null - */ - init_image?: string | null; - /** - * @description The VAE used for decoding, if the main model's default was not used - * @default null - */ - vae?: components["schemas"]["ModelIdentifierField"] | null; - /** - * @description The Qwen3 text encoder model used for Z-Image inference - * @default null - */ - qwen3_encoder?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Hrf Enabled - * @description Whether or not high resolution fix was enabled. - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - hrf_enabled?: boolean | null; + path: string; /** - * Hrf Method - * @description The high resolution fix upscale method. - * @default null + * File Size + * @description The size of the model in bytes. */ - hrf_method?: string | null; + file_size: number; /** - * Hrf Strength - * @description The high resolution fix img2img strength used in the upscale pass. - * @default null + * Name + * @description Name of the model. */ - hrf_strength?: number | null; + name: string; /** - * Positive Style Prompt - * @description The positive style prompt parameter - * @default null + * Description + * @description Model description */ - positive_style_prompt?: string | null; + description: string | null; /** - * Negative Style Prompt - * @description The negative style prompt parameter - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - negative_style_prompt?: string | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * @description The SDXL Refiner model used - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - refiner_model?: components["schemas"]["ModelIdentifierField"] | null; + source_api_response: string | null; /** - * Refiner Cfg Scale - * @description The classifier-free guidance scale parameter used for the refiner - * @default null + * Cover Image + * @description Url for image to preview model */ - refiner_cfg_scale?: number | null; + cover_image: string | null; /** - * Refiner Steps - * @description The number of steps used for the refiner - * @default null + * Type + * @default flux_redux + * @constant */ - refiner_steps?: number | null; + type: "flux_redux"; /** - * Refiner Scheduler - * @description The scheduler used for the refiner - * @default null + * Format + * @default checkpoint + * @constant */ - refiner_scheduler?: string | null; + format: "checkpoint"; /** - * Refiner Positive Aesthetic Score - * @description The aesthetic score used for the refiner - * @default null + * Base + * @default flux + * @constant */ - refiner_positive_aesthetic_score?: number | null; + base: "flux"; + }; + /** + * FaceIdentifier + * @description Outputs an image with detected face IDs printed on each face. For use with other FaceTools. + */ + FaceIdentifierInvocation: { /** - * Refiner Negative Aesthetic Score - * @description The aesthetic score used for the refiner + * @description The board to save the image to * @default null */ - refiner_negative_aesthetic_score?: number | null; + board?: components["schemas"]["BoardField"] | null; /** - * Refiner Start - * @description The start value used for refiner denoising + * @description Optional metadata to be saved with the image * @default null */ - refiner_start?: number | null; - /** - * type - * @default core_metadata - * @constant - */ - type: "core_metadata"; - } & { - [key: string]: unknown; - }; - /** - * Create Denoise Mask - * @description Creates mask for denoising model run. - */ - CreateDenoiseMaskInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9225,44 +8170,39 @@ export type components = { */ use_cache?: boolean; /** - * @description VAE - * @default null - */ - vae?: components["schemas"]["VAEField"] | null; - /** - * @description Image which will be masked + * @description Image to face detect * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description The mask to use when pasting - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Tiled - * @description Processing using overlapping tiles (reduce memory consumption) - * @default false + * Minimum Confidence + * @description Minimum confidence for face detection (lower if detection is failing) + * @default 0.5 */ - tiled?: boolean; + minimum_confidence?: number; /** - * Fp32 - * @description Whether or not to use full float32 precision + * Chunk + * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. * @default false */ - fp32?: boolean; + chunk?: boolean; /** * type - * @default create_denoise_mask + * @default face_identifier * @constant */ - type: "create_denoise_mask"; + type: "face_identifier"; }; /** - * Create Gradient Mask - * @description Creates mask for denoising. + * FaceMask + * @description Face mask creation using mediapipe face detection */ - CreateGradientMaskInvocation: { + FaceMaskInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9281,75 +8221,84 @@ export type components = { */ use_cache?: boolean; /** - * @description Image which will be masked + * @description Image to face detect * @default null */ - mask?: components["schemas"]["ImageField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Edge Radius - * @description How far to expand the edges of the mask - * @default 16 + * Face Ids + * @description Comma-separated list of face ids to mask eg '0,2,7'. Numbered from 0. Leave empty to mask all. Find face IDs with FaceIdentifier node. + * @default */ - edge_radius?: number; + face_ids?: string; /** - * Coherence Mode - * @default Gaussian Blur - * @enum {string} + * Minimum Confidence + * @description Minimum confidence for face detection (lower if detection is failing) + * @default 0.5 */ - coherence_mode?: "Gaussian Blur" | "Box Blur" | "Staged"; + minimum_confidence?: number; /** - * Minimum Denoise - * @description Minimum denoise level for the coherence region + * X Offset + * @description Offset for the X-axis of the face mask * @default 0 */ - minimum_denoise?: number; - /** - * [OPTIONAL] Image - * @description OPTIONAL: Only connect for specialized Inpainting models, masked_latents will be generated from the image with the VAE - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * [OPTIONAL] UNet - * @description OPTIONAL: If the Unet is a specialized Inpainting model, masked_latents will be generated from the image with the VAE - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; + x_offset?: number; /** - * [OPTIONAL] VAE - * @description OPTIONAL: Only connect for specialized Inpainting models, masked_latents will be generated from the image with the VAE - * @default null + * Y Offset + * @description Offset for the Y-axis of the face mask + * @default 0 */ - vae?: components["schemas"]["VAEField"] | null; + y_offset?: number; /** - * Tiled - * @description Processing using overlapping tiles (reduce memory consumption) + * Chunk + * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. * @default false */ - tiled?: boolean; + chunk?: boolean; /** - * Fp32 - * @description Whether or not to use full float32 precision + * Invert Mask + * @description Toggle to invert the mask * @default false */ - fp32?: boolean; + invert_mask?: boolean; /** * type - * @default create_gradient_mask + * @default face_mask_detection * @constant */ - type: "create_gradient_mask"; + type: "face_mask_detection"; }; /** - * Crop Image to Bounding Box - * @description Crop an image to the given bounding box. If the bounding box is omitted, the image is cropped to the non-transparent pixels. + * FaceMaskOutput + * @description Base class for FaceMask output */ - CropImageToBoundingBoxInvocation: { + FaceMaskOutput: { + /** @description The output image */ + image: components["schemas"]["ImageField"]; /** - * @description The board to save the image to - * @default null + * Width + * @description The width of the image in pixels */ - board?: components["schemas"]["BoardField"] | null; + width: number; + /** + * Height + * @description The height of the image in pixels + */ + height: number; + /** + * type + * @default face_mask_output + * @constant + */ + type: "face_mask_output"; + /** @description The output mask */ + mask: components["schemas"]["ImageField"]; + }; + /** + * FaceOff + * @description Bound, extract, and mask a face from an image using MediaPipe detection + */ + FaceOffInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -9373,86 +8322,112 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to crop + * @description Image for face detection * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description The bounding box to crop the image to - * @default null - */ - bounding_box?: components["schemas"]["BoundingBoxField"] | null; - /** - * type - * @default crop_image_to_bounding_box - * @constant + * Face Id + * @description The face ID to process, numbered from 0. Multiple faces not supported. Find a face's ID with FaceIdentifier node. + * @default 0 */ - type: "crop_image_to_bounding_box"; - }; - /** - * Crop Latents - * @description Crops a latent-space tensor to a box specified in image-space. The box dimensions and coordinates must be - * divisible by the latent scale factor of 8. - */ - CropLatentsCoreInvocation: { + face_id?: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Minimum Confidence + * @description Minimum confidence for face detection (lower if detection is failing) + * @default 0.5 */ - id: string; + minimum_confidence?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * X Offset + * @description X-axis offset of the mask + * @default 0 */ - is_intermediate?: boolean; + x_offset?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Y Offset + * @description Y-axis offset of the mask + * @default 0 */ - use_cache?: boolean; + y_offset?: number; /** - * @description Latents tensor - * @default null + * Padding + * @description All-axis padding around the mask in pixels + * @default 0 */ - latents?: components["schemas"]["LatentsField"] | null; + padding?: number; /** - * X - * @description The left x coordinate (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. - * @default null + * Chunk + * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. + * @default false */ - x?: number | null; + chunk?: boolean; /** - * Y - * @description The top y coordinate (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. - * @default null + * type + * @default face_off + * @constant */ - y?: number | null; + type: "face_off"; + }; + /** + * FaceOffOutput + * @description Base class for FaceOff Output + */ + FaceOffOutput: { + /** @description The output image */ + image: components["schemas"]["ImageField"]; /** * Width - * @description The width (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. - * @default null + * @description The width of the image in pixels */ - width?: number | null; + width: number; /** * Height - * @description The height (in px) of the crop rectangle in image space. This value will be converted to a dimension in latent space. - * @default null + * @description The height of the image in pixels */ - height?: number | null; + height: number; /** * type - * @default crop_latents + * @default face_off_output * @constant */ - type: "crop_latents"; + type: "face_off_output"; + /** @description The output mask */ + mask: components["schemas"]["ImageField"]; + /** + * X + * @description The x coordinate of the bounding box's left side + */ + x: number; + /** + * Y + * @description The y coordinate of the bounding box's top side + */ + y: number; }; /** - * Crop Latents - * @description Crops latents + * FieldKind + * @description The kind of field. + * - `Input`: An input field on a node. + * - `Output`: An output field on a node. + * - `Internal`: A field which is treated as an input, but cannot be used in node definitions. Metadata is + * one example. It is provided to nodes via the WithMetadata class, and we want to reserve the field name + * "metadata" for this on all nodes. `FieldKind` is used to short-circuit the field name validation logic, + * allowing "metadata" for that field. + * - `NodeAttribute`: The field is a node attribute. These are fields which are not inputs or outputs, + * but which are used to store information about the node. For example, the `id` and `type` fields are node + * attributes. + * + * The presence of this in `json_schema_extra["field_kind"]` is used when initializing node schemas on app + * startup, and when generating the OpenAPI schema for the workflow editor. + * @enum {string} */ - CropLatentsInvocation: { + FieldKind: "input" | "output" | "internal" | "node_attribute"; + /** + * Float Batch + * @description Create a batched generation, where the workflow is executed once for each float in the batch. + */ + FloatBatchInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9471,46 +8446,30 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * Width - * @description Width of output (px) - * @default null - */ - width?: number | null; - /** - * Height - * @description Width of output (px) - * @default null - */ - height?: number | null; - /** - * X Offset - * @description x-coordinate - * @default null + * Batch Group + * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. + * @default None + * @enum {string} */ - x_offset?: number | null; + batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; /** - * Y Offset - * @description y-coordinate + * Floats + * @description The floats to batch over * @default null */ - y_offset?: number | null; + floats?: number[] | null; /** * type - * @default lcrop + * @default float_batch * @constant */ - type: "lcrop"; + type: "float_batch"; }; /** - * Crossover Prompt - * @description Performs a crossover operation on two parent seed vectors to generate a new prompt. + * Float Collection Primitive + * @description A collection of float primitive values */ - CrossoverPromptInvocation: { + FloatCollectionInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9525,87 +8484,97 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Resolutions Dict - * @description Private field for id substitutions dict cache - * @default {} - */ - resolutions_dict?: { - [key: string]: unknown; - }; - /** - * Lookups - * @description Lookup table(s) containing template(s) (JSON) + * Collection + * @description The collection of float values * @default [] */ - lookups?: string | string[]; + collection?: number[]; /** - * Remove Negatives - * @description Whether to strip out text between [] - * @default false + * type + * @default float_collection + * @constant */ - remove_negatives?: boolean; + type: "float_collection"; + }; + /** + * FloatCollectionOutput + * @description Base class for nodes that output a collection of floats + */ + FloatCollectionOutput: { /** - * Strip Parens Probability - * @description Probability of removing attention group weightings - * @default 0 + * Collection + * @description The float collection */ - strip_parens_probability?: number; + collection: number[]; /** - * Resolutions - * @description JSON structure of substitutions by id by tag - * @default [] + * type + * @default float_collection_output + * @constant */ - resolutions?: string | string[]; + type: "float_collection_output"; + }; + /** + * Float Generator + * @description Generated a range of floats for use in a batched generation + */ + FloatGenerator: { /** - * Parent A Seed Vector In - * @description JSON array of seeds for Parent A's generation - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - parent_a_seed_vector_in?: string | null; + id: string; /** - * Parent B Seed Vector In - * @description JSON array of seeds for Parent B's generation - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - parent_b_seed_vector_in?: string | null; + is_intermediate?: boolean; /** - * Child A Or B - * @description True for Child A (Parent A + B's branch), False for Child B (Parent B + A's branch) + * Use Cache + * @description Whether or not to use the cache * @default true */ - child_a_or_b?: boolean; + use_cache?: boolean; /** - * Crossover Non Terminal - * @description Optional: The non-terminal (key in lookups) to target for the crossover branch. If None, a random one will be chosen from available non-terminals. - * @default null + * Generator Type + * @description The float generator. */ - crossover_non_terminal?: string | null; + generator: components["schemas"]["FloatGeneratorField"]; /** * type - * @default crossover_prompt + * @default float_generator * @constant */ - type: "crossover_prompt"; + type: "float_generator"; }; + /** FloatGeneratorField */ + FloatGeneratorField: Record; /** - * OpenCV Inpaint - * @description Simple inpaint using opencv. + * FloatGeneratorOutput + * @description Base class for nodes that output a collection of floats */ - CvInpaintInvocation: { + FloatGeneratorOutput: { /** - * @description The board to save the image to - * @default null + * Floats + * @description The generated floats */ - board?: components["schemas"]["BoardField"] | null; + floats: number[]; /** - * @description Optional metadata to be saved with the image - * @default null + * type + * @default float_generator_output + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "float_generator_output"; + }; + /** + * Float Primitive + * @description A float primitive value + */ + FloatInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9624,37 +8593,23 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to inpaint - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The mask to use when inpainting - * @default null + * Value + * @description The float value + * @default 0 */ - mask?: components["schemas"]["ImageField"] | null; + value?: number; /** * type - * @default cv_inpaint + * @default float * @constant */ - type: "cv_inpaint"; + type: "float"; }; /** - * DW Openpose Detection - * @description Generates an openpose pose from an image using DWPose + * Float Range + * @description Creates a range */ - DWOpenposeDetectionInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FloatLinearRangeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9673,37 +8628,35 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Draw Body - * @default true + * Start + * @description The first value of the range + * @default 5 */ - draw_body?: boolean; + start?: number; /** - * Draw Face - * @default false + * Stop + * @description The last value of the range + * @default 10 */ - draw_face?: boolean; + stop?: number; /** - * Draw Hands - * @default false + * Steps + * @description number of values to interpolate over (including start and stop) + * @default 30 */ - draw_hands?: boolean; + steps?: number; /** * type - * @default dw_openpose_detection + * @default float_range * @constant */ - type: "dw_openpose_detection"; + type: "float_range"; }; /** - * Decode Invisible Watermark - * @description Decode an invisible watermark from an image. + * Float Math + * @description Performs floating point math. */ - DecodeInvisibleWatermarkInvocation: { + FloatMathInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9722,28 +8675,53 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to decode the watermark from - * @default null + * Operation + * @description The operation to perform + * @default ADD + * @enum {string} */ - image?: components["schemas"]["ImageField"] | null; + operation?: "ADD" | "SUB" | "MUL" | "DIV" | "EXP" | "ABS" | "SQRT" | "MIN" | "MAX"; /** - * Length - * @description The expected watermark length in bytes - * @default 8 + * A + * @description The first number + * @default 1 */ - length?: number; + a?: number; + /** + * B + * @description The second number + * @default 1 + */ + b?: number; /** * type - * @default decode_watermark + * @default float_math * @constant */ - type: "decode_watermark"; + type: "float_math"; + }; + /** + * FloatOutput + * @description Base class for nodes that output a single float + */ + FloatOutput: { + /** + * Value + * @description The output float + */ + value: number; + /** + * type + * @default float_output + * @constant + */ + type: "float_output"; }; /** - * Default XYImage Tile Generator - * @description Cuts up an image into overlapping tiles and outputs a string representation of the tiles to use + * Float To Integer + * @description Rounds a float number to (a multiple of) an integer. */ - DefaultXYTileGenerator: { + FloatToIntegerInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9762,128 +8740,160 @@ export type components = { */ use_cache?: boolean; /** - * @description The input image - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Tile Width - * @description x resolution of generation tile (must be a multiple of 8) - * @default 576 - */ - tile_width?: number; - /** - * Tile Height - * @description y resolution of generation tile (must be a multiple of 8) - * @default 576 + * Value + * @description The value to round + * @default 0 */ - tile_height?: number; + value?: number; /** - * Overlap - * @description tile overlap size (must be a multiple of 8) - * @default 128 + * Multiple of + * @description The multiple to round to + * @default 1 */ - overlap?: number; + multiple?: number; /** - * Adjust Tile Size - * @description adjust tile size to account for overlap - * @default true + * Method + * @description The method to use for rounding + * @default Nearest + * @enum {string} */ - adjust_tile_size?: boolean; + method?: "Nearest" | "Floor" | "Ceiling" | "Truncate"; /** * type - * @default default_xy_tile_generator + * @default float_to_int * @constant */ - type: "default_xy_tile_generator"; + type: "float_to_int"; }; /** - * DeleteAllExceptCurrentResult - * @description Result of deleting all except current + * FLUX2 Denoise + * @description Run denoising process with a FLUX.2 Klein transformer model. + * + * This node is designed for FLUX.2 Klein models which use Qwen3 as the text encoder. + * It does not support ControlNet, IP-Adapters, or regional prompting. */ - DeleteAllExceptCurrentResult: { + Flux2DenoiseInvocation: { /** - * Deleted - * @description Number of queue items deleted + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - deleted: number; - }; - /** DeleteBoardResult */ - DeleteBoardResult: { + id: string; /** - * Board Id - * @description The id of the board that was deleted. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - board_id: string; + is_intermediate?: boolean; /** - * Deleted Board Images - * @description The image names of the board-images relationships that were deleted. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - deleted_board_images: string[]; + use_cache?: boolean; /** - * Deleted Images - * @description The names of the images that were deleted. + * @description Latents tensor + * @default null */ - deleted_images: string[]; - }; - /** - * DeleteByDestinationResult - * @description Result of deleting by a destination - */ - DeleteByDestinationResult: { + latents?: components["schemas"]["LatentsField"] | null; /** - * Deleted - * @description Number of queue items deleted + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - deleted: number; - }; - /** DeleteImagesResult */ - DeleteImagesResult: { + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Affected Boards - * @description The ids of boards affected by the delete operation + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - affected_boards: string[]; + denoising_start?: number; /** - * Deleted Images - * @description The names of the images that were deleted + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - deleted_images: string[]; - }; - /** - * DeleteOrphanedModelsRequest - * @description Request to delete specific orphaned model directories. - */ - DeleteOrphanedModelsRequest: { + denoising_end?: number; /** - * Paths - * @description List of relative paths to delete + * Add Noise + * @description Add noise based on denoising start. + * @default true */ - paths: string[]; - }; - /** - * DeleteOrphanedModelsResponse - * @description Response from deleting orphaned models. - */ - DeleteOrphanedModelsResponse: { + add_noise?: boolean; /** - * Deleted - * @description Paths that were successfully deleted + * Transformer + * @description Flux model (Transformer) to load + * @default null */ - deleted: string[]; + transformer?: components["schemas"]["TransformerField"] | null; /** - * Errors - * @description Paths that had errors, with error messages + * @description Positive conditioning tensor + * @default null */ - errors: { - [key: string]: string; - }; + positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | null; + /** + * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. + * @default null + */ + negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | null; + /** + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 1 + */ + cfg_scale?: number; + /** + * Width + * @description Width of the generated image. + * @default 1024 + */ + width?: number; + /** + * Height + * @description Height of the generated image. + * @default 1024 + */ + height?: number; + /** + * Num Steps + * @description Number of diffusion steps. Use 4 for distilled models, 28+ for base models. + * @default 4 + */ + num_steps?: number; + /** + * Scheduler + * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. + * @default euler + * @enum {string} + */ + scheduler?: "euler" | "heun" | "lcm"; + /** + * Seed + * @description Randomness seed for reproducibility. + * @default 0 + */ + seed?: number; + /** + * @description FLUX.2 VAE model (required for BN statistics). + * @default null + */ + vae?: components["schemas"]["VAEField"] | null; + /** + * Reference Images + * @description FLUX Kontext conditioning (reference images for multi-reference image editing). + * @default null + */ + kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; + /** + * type + * @default flux2_denoise + * @constant + */ + type: "flux2_denoise"; }; /** - * Denoise - SD1.5, SDXL - * @description Denoises noisy latents to decodable images + * Apply LoRA Collection - Flux2 Klein + * @description Applies a collection of LoRAs to a FLUX.2 Klein transformer and/or Qwen3 text encoder. */ - DenoiseLatentsInvocation: { + Flux2KleinLoRACollectionLoader: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -9902,106 +8912,121 @@ export type components = { */ use_cache?: boolean; /** - * Positive Conditioning - * @description Positive conditioning tensor + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. * @default null */ - positive_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * Negative Conditioning - * @description Negative conditioning tensor + * Transformer + * @description Transformer * @default null */ - negative_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; + transformer?: components["schemas"]["TransformerField"] | null; /** - * @description Noise tensor + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder * @default null */ - noise?: components["schemas"]["LatentsField"] | null; - /** - * Steps - * @description Number of steps to run - * @default 10 - */ - steps?: number; + qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 7.5 + * type + * @default flux2_klein_lora_collection_loader + * @constant */ - cfg_scale?: number | number[]; + type: "flux2_klein_lora_collection_loader"; + }; + /** + * Apply LoRA - Flux2 Klein + * @description Apply a LoRA model to a FLUX.2 Klein transformer and/or Qwen3 text encoder. + */ + Flux2KleinLoRALoaderInvocation: { /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - denoising_start?: number; + id: string; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - denoising_end?: number; + is_intermediate?: boolean; /** - * Scheduler - * @description Scheduler to use during inference - * @default euler - * @enum {string} + * Use Cache + * @description Whether or not to use the cache + * @default true */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + use_cache?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) + * LoRA + * @description LoRA model to load * @default null */ - unet?: components["schemas"]["UNetField"] | null; + lora?: components["schemas"]["ModelIdentifierField"] | null; /** - * Control - * @default null + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 */ - control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + weight?: number; /** - * IP-Adapter - * @description IP-Adapter to apply + * Transformer + * @description Transformer * @default null */ - ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + transformer?: components["schemas"]["TransformerField"] | null; /** - * T2I-Adapter - * @description T2I-Adapter(s) to apply + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder * @default null */ - t2i_adapter?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; /** - * CFG Rescale Multiplier - * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR - * @default 0 + * type + * @default flux2_klein_lora_loader + * @constant */ - cfg_rescale_multiplier?: number; + type: "flux2_klein_lora_loader"; + }; + /** + * Flux2KleinLoRALoaderOutput + * @description FLUX.2 Klein LoRA Loader Output + */ + Flux2KleinLoRALoaderOutput: { /** - * @description Latents tensor + * Transformer + * @description Transformer * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + transformer: components["schemas"]["TransformerField"] | null; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder * @default null */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + qwen3_encoder: components["schemas"]["Qwen3EncoderField"] | null; /** * type - * @default denoise_latents + * @default flux2_klein_lora_loader_output * @constant */ - type: "denoise_latents"; + type: "flux2_klein_lora_loader_output"; }; - /** Denoise - SD1.5, SDXL + Metadata */ - DenoiseLatentsMetaInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + /** + * Main Model - Flux2 Klein + * @description Loads a Flux2 Klein model, outputting its submodels. + * + * Flux2 Klein uses Qwen3 as the text encoder instead of CLIP+T5. + * It uses a 32-channel VAE (AutoencoderKLFlux2) instead of the 16-channel FLUX.1 VAE. + * + * When using a Diffusers format model, both VAE and Qwen3 encoder are extracted + * automatically from the main model. You can override with standalone models: + * - Transformer: Always from Flux2 Klein main model + * - VAE: From main model (Diffusers) or standalone VAE + * - Qwen3 Encoder: From main model (Diffusers) or standalone Qwen3 model + */ + Flux2KleinModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10020,141 +9045,137 @@ export type components = { */ use_cache?: boolean; /** - * Positive Conditioning - * @description Positive conditioning tensor - * @default null + * Transformer + * @description Flux model (Transformer) to load */ - positive_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; + model: components["schemas"]["ModelIdentifierField"]; /** - * Negative Conditioning - * @description Negative conditioning tensor - * @default null + * VAE + * @description Standalone VAE model. Flux2 Klein uses the same VAE as FLUX (16-channel). If not provided, VAE will be loaded from the Qwen3 Source model. + * @default null */ - negative_conditioning?: components["schemas"]["ConditioningField"] | components["schemas"]["ConditioningField"][] | null; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * @description Noise tensor + * Qwen3 Encoder + * @description Standalone Qwen3 Encoder model. If not provided, encoder will be loaded from the Qwen3 Source model. * @default null */ - noise?: components["schemas"]["LatentsField"] | null; + qwen3_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Steps - * @description Number of steps to run - * @default 10 + * Qwen3 Source (Diffusers) + * @description Diffusers Flux2 Klein model to extract VAE and/or Qwen3 encoder from. Use this if you don't have separate VAE/Qwen3 models. Ignored if both VAE and Qwen3 Encoder are provided separately. + * @default null */ - steps?: number; + qwen3_source_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 7.5 + * Max Seq Length + * @description Max sequence length for the Qwen3 encoder. + * @default 512 + * @enum {integer} */ - cfg_scale?: number | number[]; + max_seq_len?: 256 | 512; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * type + * @default flux2_klein_model_loader + * @constant */ - denoising_start?: number; + type: "flux2_klein_model_loader"; + }; + /** + * Flux2KleinModelLoaderOutput + * @description Flux2 Klein model loader output. + */ + Flux2KleinModelLoaderOutput: { /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * Transformer + * @description Transformer */ - denoising_end?: number; + transformer: components["schemas"]["TransformerField"]; /** - * Scheduler - * @description Scheduler to use during inference - * @default euler - * @enum {string} + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + qwen3_encoder: components["schemas"]["Qwen3EncoderField"]; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * VAE + * @description VAE */ - unet?: components["schemas"]["UNetField"] | null; + vae: components["schemas"]["VAEField"]; /** - * Control - * @default null + * Max Seq Length + * @description The max sequence length for the Qwen3 encoder. + * @enum {integer} */ - control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + max_seq_len: 256 | 512; /** - * IP-Adapter - * @description IP-Adapter to apply - * @default null + * type + * @default flux2_klein_model_loader_output + * @constant */ - ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + type: "flux2_klein_model_loader_output"; + }; + /** + * Prompt - Flux2 Klein + * @description Encodes and preps a prompt for Flux2 Klein image generation. + * + * Flux2 Klein uses Qwen3 as the text encoder, extracting hidden states from + * layers (9, 18, 27) and stacking them for richer text representations. + * This matches the diffusers Flux2KleinPipeline implementation exactly. + */ + Flux2KleinTextEncoderInvocation: { /** - * T2I-Adapter - * @description T2I-Adapter(s) to apply - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - t2i_adapter?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + id: string; /** - * CFG Rescale Multiplier - * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR - * @default 0 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cfg_rescale_multiplier?: number; + is_intermediate?: boolean; /** - * @description Latents tensor - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - latents?: components["schemas"]["LatentsField"] | null; + use_cache?: boolean; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * Prompt + * @description Text prompt to encode. * @default null */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + prompt?: string | null; /** - * type - * @default denoise_latents_meta - * @constant + * Qwen3 Encoder + * @description Qwen3 tokenizer and text encoder + * @default null */ - type: "denoise_latents_meta"; - }; - /** - * DenoiseMaskField - * @description An inpaint mask field - */ - DenoiseMaskField: { + qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; /** - * Mask Name - * @description The name of the mask image + * Max Seq Len + * @description Max sequence length for the Qwen3 encoder. + * @default 512 + * @enum {integer} */ - mask_name: string; + max_seq_len?: 256 | 512; /** - * Masked Latents Name - * @description The name of the masked image latents + * @description A mask defining the region that this conditioning prompt applies to. * @default null */ - masked_latents_name?: string | null; - /** - * Gradient - * @description Used for gradient inpainting - * @default false - */ - gradient?: boolean; - }; - /** - * DenoiseMaskOutput - * @description Base class for nodes that output a single image - */ - DenoiseMaskOutput: { - /** @description Mask for denoise model run */ - denoise_mask: components["schemas"]["DenoiseMaskField"]; + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default denoise_mask_output + * @default flux2_klein_text_encoder * @constant */ - type: "denoise_mask_output"; + type: "flux2_klein_text_encoder"; }; /** - * Depth Anything Depth Estimation - * @description Generates a depth map using a Depth Anything model. + * Latents to Image - FLUX2 + * @description Generates an image from latents using FLUX.2 Klein's 32-channel VAE. */ - DepthAnythingDepthEstimationInvocation: { + Flux2VaeDecodeInvocation: { /** * @description The board to save the image to * @default null @@ -10183,29 +9204,27 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to process + * @description Latents tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * Model Size - * @description The size of the depth model to use - * @default small_v2 - * @enum {string} + * @description VAE + * @default null */ - model_size?: "large" | "base" | "small" | "small_v2"; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default depth_anything_depth_estimation + * @default flux2_vae_decode * @constant */ - type: "depth_anything_depth_estimation"; + type: "flux2_vae_decode"; }; /** - * Divide Integers - * @description Divides two numbers + * Image to Latents - FLUX2 + * @description Encodes an image into latents using FLUX.2 Klein's 32-channel VAE. */ - DivideInvocation: { + Flux2VaeEncodeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10224,295 +9243,266 @@ export type components = { */ use_cache?: boolean; /** - * A - * @description The first number - * @default 0 + * @description The image to encode. + * @default null */ - a?: number; + image?: components["schemas"]["ImageField"] | null; /** - * B - * @description The second number - * @default 0 + * @description VAE + * @default null */ - b?: number; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default div + * @default flux2_vae_encode * @constant */ - type: "div"; + type: "flux2_vae_encode"; }; /** - * DownloadCancelledEvent - * @description Event model for download_cancelled + * Flux2VariantType + * @description FLUX.2 model variants. + * @enum {string} */ - DownloadCancelledEvent: { + Flux2VariantType: "klein_4b" | "klein_9b" | "klein_9b_base"; + /** + * FluxConditioningCollectionOutput + * @description Base class for nodes that output a collection of conditioning tensors + */ + FluxConditioningCollectionOutput: { /** - * Timestamp - * @description The timestamp of the event + * Collection + * @description The output conditioning tensors */ - timestamp: number; + collection: components["schemas"]["FluxConditioningField"][]; /** - * Source - * @description The source of the download + * type + * @default flux_conditioning_collection_output + * @constant */ - source: string; + type: "flux_conditioning_collection_output"; }; /** - * DownloadCompleteEvent - * @description Event model for download_complete + * FluxConditioningField + * @description A conditioning tensor primitive value */ - DownloadCompleteEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; + FluxConditioningField: { /** - * Source - * @description The source of the download + * Conditioning Name + * @description The name of conditioning tensor */ - source: string; + conditioning_name: string; /** - * Download Path - * @description The local path where the download is saved + * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. + * @default null */ - download_path: string; - /** - * Total Bytes - * @description The total number of bytes downloaded - */ - total_bytes: number; + mask?: components["schemas"]["TensorField"] | null; }; /** - * DownloadErrorEvent - * @description Event model for download_error + * FluxConditioningOutput + * @description Base class for nodes that output a single conditioning tensor */ - DownloadErrorEvent: { + FluxConditioningOutput: { + /** @description Conditioning tensor */ + conditioning: components["schemas"]["FluxConditioningField"]; /** - * Timestamp - * @description The timestamp of the event + * type + * @default flux_conditioning_output + * @constant */ - timestamp: number; + type: "flux_conditioning_output"; + }; + /** + * Control LoRA - FLUX + * @description LoRA model and Image to use with FLUX transformer generation. + */ + FluxControlLoRALoaderInvocation: { /** - * Source - * @description The source of the download + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; + id: string; /** - * Error Type - * @description The type of error + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - error_type: string; + is_intermediate?: boolean; /** - * Error - * @description The error message + * Use Cache + * @description Whether or not to use the cache + * @default true */ - error: string; - }; - /** - * DownloadJob - * @description Class to monitor and control a model download request. - */ - DownloadJob: { + use_cache?: boolean; /** - * Id - * @description Numeric ID of this job - * @default -1 + * Control LoRA + * @description Control LoRA model to load + * @default null */ - id?: number; + lora?: components["schemas"]["ModelIdentifierField"] | null; /** - * Dest - * Format: path - * @description Initial destination of downloaded model on local disk; a directory or file path + * @description The image to encode. + * @default null */ - dest: string; + image?: components["schemas"]["ImageField"] | null; /** - * Download Path - * @description Final location of downloaded file or directory + * Weight + * @description The weight of the LoRA. + * @default 1 */ - download_path?: string | null; + weight?: number; /** - * @description Status of the download - * @default waiting + * type + * @default flux_control_lora_loader + * @constant */ - status?: components["schemas"]["DownloadJobStatus"]; + type: "flux_control_lora_loader"; + }; + /** + * FluxControlLoRALoaderOutput + * @description Flux Control LoRA Loader Output + */ + FluxControlLoRALoaderOutput: { /** - * Bytes - * @description Bytes downloaded so far - * @default 0 + * Flux Control LoRA + * @description Control LoRAs to apply on model loading + * @default null */ - bytes?: number; + control_lora: components["schemas"]["ControlLoRAField"]; /** - * Total Bytes - * @description Total file size (bytes) - * @default 0 + * type + * @default flux_control_lora_loader_output + * @constant */ - total_bytes?: number; + type: "flux_control_lora_loader_output"; + }; + /** FluxControlNetField */ + FluxControlNetField: { + /** @description The control image */ + image: components["schemas"]["ImageField"]; + /** @description The ControlNet model to use */ + control_model: components["schemas"]["ModelIdentifierField"]; /** - * Error Type - * @description Name of exception that caused an error + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - error_type?: string | null; + control_weight?: number | number[]; /** - * Error - * @description Traceback of the exception that caused an error + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - error?: string | null; + begin_step_percent?: number; /** - * Source - * Format: uri - * @description Where to download from. Specific types specified in child classes. + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - source: string; + end_step_percent?: number; /** - * Access Token - * @description authorization token for protected resources + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} */ - access_token?: string | null; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** - * Priority - * @description Queue priority; lower values are higher priority - * @default 10 + * Instantx Control Mode + * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. + * @default -1 */ - priority?: number; + instantx_control_mode?: number | null; + }; + /** + * FLUX ControlNet + * @description Collect FLUX ControlNet info to pass to other nodes. + */ + FluxControlNetInvocation: { /** - * Job Started - * @description Timestamp for when the download job started + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - job_started?: string | null; + id: string; /** - * Job Ended - * @description Timestamp for when the download job ende1d (completed or errored) + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - job_ended?: string | null; + is_intermediate?: boolean; /** - * Content Type - * @description Content type of downloaded file + * Use Cache + * @description Whether or not to use the cache + * @default true */ - content_type?: string | null; + use_cache?: boolean; /** - * Canonical Url - * @description Canonical URL to request on resume + * @description The control image + * @default null */ - canonical_url?: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Etag - * @description ETag from the remote server, if available + * @description ControlNet model to load + * @default null */ - etag?: string | null; + control_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Last Modified - * @description Last-Modified from the remote server, if available + * Control Weight + * @description The weight given to the ControlNet + * @default 1 */ - last_modified?: string | null; + control_weight?: number | number[]; /** - * Final Url - * @description Final resolved URL after redirects, if available + * Begin Step Percent + * @description When the ControlNet is first applied (% of total steps) + * @default 0 */ - final_url?: string | null; + begin_step_percent?: number; /** - * Expected Total Bytes - * @description Expected total size of the download + * End Step Percent + * @description When the ControlNet is last applied (% of total steps) + * @default 1 */ - expected_total_bytes?: number | null; + end_step_percent?: number; /** - * Resume Required - * @description True if server refused resume; restart required - * @default false + * Resize Mode + * @description The resize mode used + * @default just_resize + * @enum {string} */ - resume_required?: boolean; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** - * Resume Message - * @description Message explaining why resume is required + * Instantx Control Mode + * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. + * @default -1 */ - resume_message?: string | null; + instantx_control_mode?: number | null; /** - * Resume From Scratch - * @description True if resume metadata existed but the partial file was missing and the download restarted from the beginning - * @default false + * type + * @default flux_controlnet + * @constant */ - resume_from_scratch?: boolean; + type: "flux_controlnet"; }; /** - * DownloadJobStatus - * @description State of a download job. - * @enum {string} - */ - DownloadJobStatus: "waiting" | "running" | "paused" | "completed" | "cancelled" | "error"; - /** - * DownloadPausedEvent - * @description Event model for download_paused + * FluxControlNetOutput + * @description FLUX ControlNet info */ - DownloadPausedEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; + FluxControlNetOutput: { + /** @description ControlNet(s) to apply */ + control: components["schemas"]["FluxControlNetField"]; /** - * Source - * @description The source of the download + * type + * @default flux_controlnet_output + * @constant */ - source: string; + type: "flux_controlnet_output"; }; /** - * DownloadProgressEvent - * @description Event model for download_progress - */ - DownloadProgressEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; - /** - * Source - * @description The source of the download - */ - source: string; - /** - * Download Path - * @description The local path where the download is saved - */ - download_path: string; - /** - * Current Bytes - * @description The number of bytes downloaded so far - */ - current_bytes: number; - /** - * Total Bytes - * @description The total number of bytes to be downloaded - */ - total_bytes: number; - }; - /** - * DownloadStartedEvent - * @description Event model for download_started - */ - DownloadStartedEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; - /** - * Source - * @description The source of the download - */ - source: string; - /** - * Download Path - * @description The local path where the download is saved - */ - download_path: string; - }; - /** - * Dynamic Prompt - * @description Parses a prompt using adieyal/dynamicprompts' random or combinatorial generator + * FLUX Denoise + * @description Run denoising process with a FLUX transformer model. */ - DynamicPromptInvocation: { + FluxDenoiseInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10527,238 +9517,186 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Prompt - * @description The prompt to parse with dynamicprompts + * @description Latents tensor * @default null */ - prompt?: string | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * Max Prompts - * @description The number of prompts to generate - * @default 1 + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - max_prompts?: number; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Combinatorial - * @description Whether to use the combinatorial generator - * @default false + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - combinatorial?: boolean; + denoising_start?: number; /** - * type - * @default dynamic_prompt - * @constant + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - type: "dynamic_prompt"; - }; - /** DynamicPromptsResponse */ - DynamicPromptsResponse: { - /** Prompts */ - prompts: string[]; - /** Error */ - error?: string | null; - }; - /** - * Upscale (RealESRGAN) - * @description Upscales an image using RealESRGAN. - */ - ESRGANInvocation: { + denoising_end?: number; /** - * @description The board to save the image to - * @default null + * Add Noise + * @description Add noise based on denoising start. + * @default true */ - board?: components["schemas"]["BoardField"] | null; + add_noise?: boolean; /** - * @description Optional metadata to be saved with the image + * Transformer + * @description Flux model (Transformer) to load * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + transformer?: components["schemas"]["TransformerField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Control LoRA + * @description Control LoRA model to load + * @default null */ - id: string; + control_lora?: components["schemas"]["ControlLoRAField"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Positive Text Conditioning + * @description Positive conditioning tensor + * @default null */ - is_intermediate?: boolean; + positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Negative Text Conditioning + * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. + * @default null */ - use_cache?: boolean; + negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * @description The input image + * Redux Conditioning + * @description FLUX Redux conditioning tensor. * @default null */ - image?: components["schemas"]["ImageField"] | null; + redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; /** - * Model Name - * @description The Real-ESRGAN model to use - * @default RealESRGAN_x4plus.pth - * @enum {string} + * @description FLUX Fill conditioning. + * @default null */ - model_name?: "RealESRGAN_x4plus.pth" | "RealESRGAN_x4plus_anime_6B.pth" | "ESRGAN_SRx4_DF2KOST_official-ff704c30.pth" | "RealESRGAN_x2plus.pth"; + fill_conditioning?: components["schemas"]["FluxFillConditioningField"] | null; /** - * Tile Size - * @description Tile size for tiled ESRGAN upscaling (0=tiling disabled) - * @default 400 + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 1 */ - tile_size?: number; + cfg_scale?: number | number[]; /** - * type - * @default esrgan - * @constant + * CFG Scale Start Step + * @description Index of the first step to apply cfg_scale. Negative indices count backwards from the the last step (e.g. a value of -1 refers to the final step). + * @default 0 */ - type: "esrgan"; - }; - /** Edge */ - Edge: { - /** @description The connection for the edge's from node and field */ - source: components["schemas"]["EdgeConnection"]; - /** @description The connection for the edge's to node and field */ - destination: components["schemas"]["EdgeConnection"]; - }; - /** EdgeConnection */ - EdgeConnection: { + cfg_scale_start_step?: number; /** - * Node Id - * @description The id of the node for this edge connection + * CFG Scale End Step + * @description Index of the last step to apply cfg_scale. Negative indices count backwards from the last step (e.g. a value of -1 refers to the final step). + * @default -1 */ - node_id: string; + cfg_scale_end_step?: number; /** - * Field - * @description The field for this connection + * Width + * @description Width of the generated image. + * @default 1024 */ - field: string; - }; - /** - * Enhance Detail - * @description Enhance Detail using guided filter - */ - EnhanceDetailInvocation: { + width?: number; /** - * @description The board to save the image to - * @default null + * Height + * @description Height of the generated image. + * @default 1024 */ - board?: components["schemas"]["BoardField"] | null; + height?: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Num Steps + * @description Number of diffusion steps. Recommended values are schnell: 4, dev: 50. + * @default 4 */ - metadata?: components["schemas"]["MetadataField"] | null; + num_steps?: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Scheduler + * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. + * @default euler + * @enum {string} */ - id: string; + scheduler?: "euler" | "heun" | "lcm"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Guidance + * @description The guidance strength. Higher values adhere more strictly to the prompt, and will produce less diverse images. FLUX dev only, ignored for schnell. + * @default 4 */ - is_intermediate?: boolean; + guidance?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Seed + * @description Randomness seed for reproducibility. + * @default 0 */ - use_cache?: boolean; + seed?: number; /** - * @description The image to detail enhance + * Control + * @description ControlNet models. * @default null */ - image?: components["schemas"]["ImageField"] | null; + control?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; /** - * Filter Radius - * @description radius of filter - * @default 2 + * @description VAE + * @default null */ - filter_radius?: number; + controlnet_vae?: components["schemas"]["VAEField"] | null; /** - * Sigma - * @description sigma - * @default 0.1 + * IP-Adapter + * @description IP-Adapter to apply + * @default null */ - sigma?: number; + ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * Denoise - * @description denoise - * @default 0.1 + * Kontext Conditioning + * @description FLUX Kontext conditioning (reference image). + * @default null */ - denoise?: number; + kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; /** - * Detail Multiplier - * @description detail multiplier - * @default 2 + * Dype Preset + * @description DyPE preset for high-resolution generation. 'auto' enables automatically for resolutions > 1536px. 'area' enables automatically based on image area. '4k' uses optimized settings for 4K output. + * @default off + * @enum {string} */ - detail_multiplier?: number; + dype_preset?: "off" | "manual" | "auto" | "area" | "4k"; /** - * type - * @default enhance_detail - * @constant - */ - type: "enhance_detail"; - }; - /** EnqueueBatchResult */ - EnqueueBatchResult: { - /** - * Queue Id - * @description The ID of the queue - */ - queue_id: string; - /** - * Enqueued - * @description The total number of queue items enqueued - */ - enqueued: number; - /** - * Requested - * @description The total number of queue items requested to be enqueued - */ - requested: number; - /** @description The batch that was enqueued */ - batch: components["schemas"]["Batch"]; - /** - * Priority - * @description The priority of the enqueued batch - */ - priority: number; - /** - * Item Ids - * @description The IDs of the queue items that were enqueued + * Dype Scale + * @description DyPE magnitude (λs). Higher values = stronger extrapolation. Only used when dype_preset is not 'off'. + * @default null */ - item_ids: number[]; - }; - /** - * EscapedOutput - * @description The input str with double quotes escaped - */ - EscapedOutput: { + dype_scale?: number | null; /** - * Prompt - * @description The input str with double quotes escaped + * Dype Exponent + * @description DyPE decay speed (λt). Controls transition from low to high frequency detail. Only used when dype_preset is not 'off'. + * @default null */ - prompt: string; + dype_exponent?: number | null; /** * type - * @default escaped_quotes_output + * @default flux_denoise * @constant */ - type: "escaped_quotes_output"; + type: "flux_denoise"; }; /** - * Quote Escaper - * @description Escapes double quotes from input strings + * FLUX Denoise + Metadata + * @description Run denoising process with a FLUX transformer model + metadata. */ - EscaperInvocation: { + FluxDenoiseLatentsMetaInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10773,210 +9711,191 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Prompt - * @description the string to escape quotes from - * @default + * @description Latents tensor + * @default null */ - prompt?: string; + latents?: components["schemas"]["LatentsField"] | null; /** - * type - * @default quote_escaper - * @constant + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * @default null */ - type: "quote_escaper"; - }; - /** - * Even Split XYImage Tile Generator - * @description Cuts up an image into a number of even sized tiles with the overlap been a percentage of the tile size and outputs a string representation of the tiles to use - */ - EvenSplitXYTileGenerator: { + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - id: string; + denoising_start?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - is_intermediate?: boolean; + denoising_end?: number; /** - * Use Cache - * @description Whether or not to use the cache + * Add Noise + * @description Add noise based on denoising start. * @default true */ - use_cache?: boolean; + add_noise?: boolean; /** - * @description The input image + * Transformer + * @description Flux model (Transformer) to load * @default null */ - image?: components["schemas"]["ImageField"] | null; + transformer?: components["schemas"]["TransformerField"] | null; /** - * Num X Tiles - * @description Number of tiles to divide image into on the x axis - * @default 2 + * Control LoRA + * @description Control LoRA model to load + * @default null */ - num_x_tiles?: number; + control_lora?: components["schemas"]["ControlLoRAField"] | null; /** - * Num Y Tiles - * @description Number of tiles to divide image into on the y axis - * @default 2 + * Positive Text Conditioning + * @description Positive conditioning tensor + * @default null */ - num_y_tiles?: number; + positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * Overlap - * @description Overlap amount of tile size (0-1) - * @default 0.25 + * Negative Text Conditioning + * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. + * @default null */ - overlap?: number; + negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; /** - * type - * @default even_split_xy_tile_generator - * @constant + * Redux Conditioning + * @description FLUX Redux conditioning tensor. + * @default null */ - type: "even_split_xy_tile_generator"; - }; - /** - * EvolutionaryPromptListOutput - * @description Output for a list of generated prompts and their seed vectors. - */ - EvolutionaryPromptListOutput: { + redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; /** - * Prompt Seed Pairs - * @description List of JSON strings, each containing [prompt, seed_vector_json] + * @description FLUX Fill conditioning. + * @default null */ - prompt_seed_pairs: string[]; + fill_conditioning?: components["schemas"]["FluxFillConditioningField"] | null; /** - * Selected Pair - * @description The index-selected prompt from the output population + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 1 */ - selected_pair: string; + cfg_scale?: number | number[]; /** - * type - * @default evolutionary_prompt_list_output - * @constant + * CFG Scale Start Step + * @description Index of the first step to apply cfg_scale. Negative indices count backwards from the the last step (e.g. a value of -1 refers to the final step). + * @default 0 */ - type: "evolutionary_prompt_list_output"; - }; - /** - * Expand Mask with Fade - * @description Expands a mask with a fade effect. The mask uses black to indicate areas to keep from the generated image and white for areas to discard. - * The mask is thresholded to create a binary mask, and then a distance transform is applied to create a fade effect. - * The fade size is specified in pixels, and the mask is expanded by that amount. The result is a mask with a smooth transition from black to white. - * If the fade size is 0, the mask is returned as-is. - */ - ExpandMaskWithFadeInvocation: { + cfg_scale_start_step?: number; /** - * @description The board to save the image to - * @default null + * CFG Scale End Step + * @description Index of the last step to apply cfg_scale. Negative indices count backwards from the last step (e.g. a value of -1 refers to the final step). + * @default -1 */ - board?: components["schemas"]["BoardField"] | null; + cfg_scale_end_step?: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Width + * @description Width of the generated image. + * @default 1024 */ - metadata?: components["schemas"]["MetadataField"] | null; + width?: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Height + * @description Height of the generated image. + * @default 1024 */ - id: string; + height?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Num Steps + * @description Number of diffusion steps. Recommended values are schnell: 4, dev: 50. + * @default 4 */ - is_intermediate?: boolean; + num_steps?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Scheduler + * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. + * @default euler + * @enum {string} */ - use_cache?: boolean; + scheduler?: "euler" | "heun" | "lcm"; /** - * @description The mask to expand - * @default null + * Guidance + * @description The guidance strength. Higher values adhere more strictly to the prompt, and will produce less diverse images. FLUX dev only, ignored for schnell. + * @default 4 */ - mask?: components["schemas"]["ImageField"] | null; + guidance?: number; /** - * Threshold - * @description The threshold for the binary mask (0-255) + * Seed + * @description Randomness seed for reproducibility. * @default 0 */ - threshold?: number; + seed?: number; /** - * Fade Size Px - * @description The size of the fade in pixels - * @default 32 + * Control + * @description ControlNet models. + * @default null */ - fade_size_px?: number; + control?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; /** - * type - * @default expand_mask_with_fade - * @constant + * @description VAE + * @default null */ - type: "expand_mask_with_fade"; - }; - /** ExposedField */ - ExposedField: { - /** Nodeid */ - nodeId: string; - /** Fieldname */ - fieldName: string; - }; - /** - * Extract Image Collection Metadata (Bool) - * @description This node extracts specified metadata values as booleans from a collection of input images. - * Values are converted to boolean: truthy values become True, falsy values (including None, empty string, 0) become False. - */ - ExtractImageCollectionMetadataBooleanInvocation: { + controlnet_vae?: components["schemas"]["VAEField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * IP-Adapter + * @description IP-Adapter to apply + * @default null */ - id: string; + ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Kontext Conditioning + * @description FLUX Kontext conditioning (reference image). + * @default null */ - is_intermediate?: boolean; + kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Dype Preset + * @description DyPE preset for high-resolution generation. 'auto' enables automatically for resolutions > 1536px. 'area' enables automatically based on image area. '4k' uses optimized settings for 4K output. + * @default off + * @enum {string} */ - use_cache?: boolean; + dype_preset?: "off" | "manual" | "auto" | "area" | "4k"; /** - * Image Collection - * @description A collection of images from which to extract metadata. + * Dype Scale + * @description DyPE magnitude (λs). Higher values = stronger extrapolation. Only used when dype_preset is not 'off'. * @default null */ - images?: components["schemas"]["ImageField"][] | null; + dype_scale?: number | null; /** - * Metadata Key - * @description Metadata key to extract values for Output. Leave empty to ignore. - * @default + * Dype Exponent + * @description DyPE decay speed (λt). Controls transition from low to high frequency detail. Only used when dype_preset is not 'off'. + * @default null */ - key?: string; + dype_exponent?: number | null; /** * type - * @default extract_image_collection_metadata_boolean + * @default flux_denoise_meta * @constant */ - type: "extract_image_collection_metadata_boolean"; + type: "flux_denoise_meta"; }; /** - * Extract Image Collection Metadata (Float) - * @description This node extracts specified metadata values as floats from a collection of input images. - * Non-float values will attempt to be converted. If conversion fails, 0.0 is used. + * FluxFillConditioningField + * @description A FLUX Fill conditioning field. */ - ExtractImageCollectionMetadataFloatInvocation: { + FluxFillConditioningField: { + /** @description The FLUX Fill reference image. */ + image: components["schemas"]["ImageField"]; + /** @description The FLUX Fill inpaint mask. */ + mask: components["schemas"]["TensorField"]; + }; + /** + * FLUX Fill Conditioning + * @description Prepare the FLUX Fill conditioning data. + */ + FluxFillInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -10995,30 +9914,44 @@ export type components = { */ use_cache?: boolean; /** - * Image Collection - * @description A collection of images from which to extract metadata. + * @description The FLUX Fill reference image. * @default null */ - images?: components["schemas"]["ImageField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Metadata Key - * @description Metadata key to extract values for Output. Leave empty to ignore. - * @default + * @description The bool inpainting mask. Excluded regions should be set to False, included regions should be set to True. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; + /** + * type + * @default flux_fill + * @constant + */ + type: "flux_fill"; + }; + /** + * FluxFillOutput + * @description The conditioning output of a FLUX Fill invocation. + */ + FluxFillOutput: { + /** + * Conditioning + * @description FLUX Redux conditioning tensor */ - key?: string; + fill_cond: components["schemas"]["FluxFillConditioningField"]; /** * type - * @default extract_image_collection_metadata_float + * @default flux_fill_output * @constant */ - type: "extract_image_collection_metadata_float"; + type: "flux_fill_output"; }; /** - * Extract Image Collection Metadata (Int) - * @description This node extracts specified metadata values as integers from a collection of input images. - * Non-integer values will attempt to be converted. If conversion fails, 0 is used. + * FLUX IP-Adapter + * @description Collects FLUX IP-Adapter info to pass to other nodes. */ - ExtractImageCollectionMetadataIntegerInvocation: { + FluxIPAdapterInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11037,29 +9970,64 @@ export type components = { */ use_cache?: boolean; /** - * Image Collection - * @description A collection of images from which to extract metadata. + * @description The IP-Adapter image prompt(s). * @default null */ - images?: components["schemas"]["ImageField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Metadata Key - * @description Metadata key to extract values for Output. Leave empty to ignore. - * @default + * IP-Adapter Model + * @description The IP-Adapter model. + * @default null + */ + ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Clip Vision Model + * @description CLIP Vision model to use. + * @default ViT-L + * @constant + */ + clip_vision_model?: "ViT-L"; + /** + * Weight + * @description The weight given to the IP-Adapter + * @default 1 + */ + weight?: number | number[]; + /** + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) + * @default 0 + */ + begin_step_percent?: number; + /** + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) + * @default 1 */ - key?: string; + end_step_percent?: number; /** * type - * @default extract_image_collection_metadata_integer + * @default flux_ip_adapter * @constant */ - type: "extract_image_collection_metadata_integer"; + type: "flux_ip_adapter"; }; /** - * Extract Image Collection Metadata (String) - * @description This node extracts specified metadata values as strings from a collection of input images. + * FLUX Kontext Image Prep + * @description Prepares an image or images for use with FLUX Kontext. The first/single image is resized to the nearest + * preferred Kontext resolution. All other images are concatenated horizontally, maintaining their aspect ratio. */ - ExtractImageCollectionMetadataStringInvocation: { + FluxKontextConcatenateImagesInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11078,29 +10046,37 @@ export type components = { */ use_cache?: boolean; /** - * Image Collection - * @description A collection of images from which to extract metadata. + * Images + * @description The images to concatenate * @default null */ images?: components["schemas"]["ImageField"][] | null; /** - * Metadata Key - * @description Metadata key to extract values for Output. Leave empty to ignore. - * @default + * Use Preferred Resolution + * @description Use FLUX preferred resolutions for the first image + * @default true */ - key?: string; + use_preferred_resolution?: boolean; /** * type - * @default extract_image_collection_metadata_string + * @default flux_kontext_image_prep * @constant */ - type: "extract_image_collection_metadata_string"; + type: "flux_kontext_image_prep"; + }; + /** + * FluxKontextConditioningField + * @description A conditioning field for FLUX Kontext (reference image). + */ + FluxKontextConditioningField: { + /** @description The Kontext reference image. */ + image: components["schemas"]["ImageField"]; }; /** - * Extrude Depth from Mask - * @description Node for creating fake depth by "extruding" a mask using OpenCV. + * Kontext Conditioning - FLUX + * @description Prepares a reference image for FLUX Kontext conditioning. */ - ExtrudeDepthInvocation: { + FluxKontextInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11115,74 +10091,43 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * @description The mask from which to extrude + * @description The Kontext reference image. * @default null */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Direction - * @description Extrude direction in degrees - * @default 45 - */ - direction?: number; - /** - * Shift - * @description Number of pixels to shift bottom from top - * @default 40 - */ - shift?: number; - /** - * Close Point - * @description Closest extrusion depth - * @default 180 - */ - close_point?: number; - /** - * Far Point - * @description Farthest extrusion depth - * @default 80 - */ - far_point?: number; - /** - * Bg Threshold - * @description Background threshold - * @default 10 - */ - bg_threshold?: number; - /** - * Bg Depth - * @description Target background depth - * @default 0 - */ - bg_depth?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Steps - * @description Number of steps in extrusion gradient - * @default 100 + * type + * @default flux_kontext + * @constant */ - steps?: number; + type: "flux_kontext"; + }; + /** + * FluxKontextOutput + * @description The conditioning output of a FLUX Kontext invocation. + */ + FluxKontextOutput: { /** - * Invert - * @description Inverts mask image before extruding - * @default false + * Kontext Conditioning + * @description FLUX Kontext conditioning (reference image) */ - invert?: boolean; + kontext_cond: components["schemas"]["FluxKontextConditioningField"]; /** * type - * @default cv_extrude_depth + * @default flux_kontext_output * @constant */ - type: "cv_extrude_depth"; + type: "flux_kontext_output"; }; /** - * FLUX Conditioning Collection Toggle - * @description Allows boolean selection between two separate FLUX conditioning collection inputs + * Apply LoRA - FLUX + * @description Apply a LoRA model to a FLUX transformer and/or text encoder. */ - FLUXConditioningCollectionToggleInvocation: { + FluxLoRALoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11201,80 +10146,77 @@ export type components = { */ use_cache?: boolean; /** - * Use Second - * @description Use 2nd Input - * @default false + * LoRA + * @description LoRA model to load + * @default null + */ + lora?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 + */ + weight?: number; + /** + * FLUX Transformer + * @description Transformer + * @default null */ - use_second?: boolean; + transformer?: components["schemas"]["TransformerField"] | null; /** - * Col1 - * @description First FLUX Conditioning Collection Input + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - col1?: components["schemas"]["FluxConditioningField"][] | null; + clip?: components["schemas"]["CLIPField"] | null; /** - * Col2 - * @description Second FLUX Conditioning Collection Input + * T5 Encoder + * @description T5 tokenizer and text encoder * @default null */ - col2?: components["schemas"]["FluxConditioningField"][] | null; + t5_encoder?: components["schemas"]["T5EncoderField"] | null; /** * type - * @default flux_conditioning_collection_toggle + * @default flux_lora_loader * @constant */ - type: "flux_conditioning_collection_toggle"; + type: "flux_lora_loader"; }; /** - * FLUX Conditioning Toggle - * @description Allows boolean selection between two separate FLUX conditioning inputs + * FluxLoRALoaderOutput + * @description FLUX LoRA Loader Output */ - FLUXConditioningToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; + FluxLoRALoaderOutput: { /** - * Use Second - * @description Use 2nd Input - * @default false + * FLUX Transformer + * @description Transformer + * @default null */ - use_second?: boolean; + transformer: components["schemas"]["TransformerField"] | null; /** - * @description First FLUX Conditioning Input + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - cond1?: components["schemas"]["FluxConditioningField"] | null; + clip: components["schemas"]["CLIPField"] | null; /** - * @description Second FLUX Conditioning Input + * T5 Encoder + * @description T5 tokenizer and text encoder * @default null */ - cond2?: components["schemas"]["FluxConditioningField"] | null; + t5_encoder: components["schemas"]["T5EncoderField"] | null; /** * type - * @default flux_conditioning_toggle + * @default flux_lora_loader_output * @constant */ - type: "flux_conditioning_toggle"; + type: "flux_lora_loader_output"; }; /** - * Apply LoRA Collection - FLUX - * @description Applies a collection of LoRAs to a FLUX transformer. + * Main Model - FLUX + * @description Loads a flux base model, outputting its submodels. */ - FLUXLoRACollectionLoader: { + FluxModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11293,122 +10235,91 @@ export type components = { */ use_cache?: boolean; /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. + * @description Flux model (Transformer) to load * @default null */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Transformer - * @description Transformer + * T5 Encoder + * @description T5 tokenizer and text encoder * @default null */ - transformer?: components["schemas"]["TransformerField"] | null; + t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * CLIP Embed + * @description CLIP Embed loader * @default null */ - clip?: components["schemas"]["CLIPField"] | null; + clip_embed_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * T5 Encoder - * @description T5 tokenizer and text encoder + * VAE + * @description VAE model to load * @default null */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; /** * type - * @default flux_lora_collection_loader + * @default flux_model_loader * @constant */ - type: "flux_lora_collection_loader"; + type: "flux_model_loader"; }; /** - * FLUXRedux_Checkpoint_Config - * @description Model config for FLUX Tools Redux model. + * FluxModelLoaderOutput + * @description Flux base model loader output */ - FLUXRedux_Checkpoint_Config: { + FluxModelLoaderOutput: { /** - * Key - * @description A unique key for this model. + * Transformer + * @description Transformer */ - key: string; + transformer: components["schemas"]["TransformerField"]; /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - source_api_response: string | null; + clip: components["schemas"]["CLIPField"]; /** - * Cover Image - * @description Url for image to preview model + * T5 Encoder + * @description T5 tokenizer and text encoder */ - cover_image: string | null; + t5_encoder: components["schemas"]["T5EncoderField"]; /** - * Type - * @default flux_redux - * @constant + * VAE + * @description VAE */ - type: "flux_redux"; + vae: components["schemas"]["VAEField"]; /** - * Format - * @default checkpoint - * @constant + * Max Seq Length + * @description The max sequence length to used for the T5 encoder. (256 for schnell transformer, 512 for dev transformer) + * @enum {integer} */ - format: "checkpoint"; + max_seq_len: 256 | 512; /** - * Base - * @default flux + * type + * @default flux_model_loader_output * @constant */ - base: "flux"; + type: "flux_model_loader_output"; }; /** - * FaceIdentifier - * @description Outputs an image with detected face IDs printed on each face. For use with other FaceTools. + * FluxReduxConditioningField + * @description A FLUX Redux conditioning tensor primitive value */ - FaceIdentifierInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + FluxReduxConditioningField: { + /** @description The Redux image conditioning tensor. */ + conditioning: components["schemas"]["TensorField"]; /** - * @description Optional metadata to be saved with the image + * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + mask?: components["schemas"]["TensorField"] | null; + }; + /** + * FLUX Redux + * @description Runs a FLUX Redux model to generate a conditioning tensor. + */ + FluxReduxInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11427,39 +10338,69 @@ export type components = { */ use_cache?: boolean; /** - * @description Image to face detect + * @description The FLUX Redux image prompt. * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Minimum Confidence - * @description Minimum confidence for face detection (lower if detection is failing) - * @default 0.5 + * @description The bool mask associated with this FLUX Redux image prompt. Excluded regions should be set to False, included regions should be set to True. + * @default null */ - minimum_confidence?: number; + mask?: components["schemas"]["TensorField"] | null; /** - * Chunk - * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. - * @default false + * FLUX Redux Model + * @description The FLUX Redux model to use. + * @default null */ - chunk?: boolean; + redux_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Downsampling Factor + * @description Redux Downsampling Factor (1-9) + * @default 1 + */ + downsampling_factor?: number; + /** + * Downsampling Function + * @description Redux Downsampling Function + * @default area + * @enum {string} + */ + downsampling_function?: "nearest" | "bilinear" | "bicubic" | "area" | "nearest-exact"; + /** + * Weight + * @description Redux weight (0.0-1.0) + * @default 1 + */ + weight?: number; /** * type - * @default face_identifier + * @default flux_redux * @constant */ - type: "face_identifier"; + type: "flux_redux"; }; /** - * FaceMask - * @description Face mask creation using mediapipe face detection + * FluxReduxOutput + * @description The conditioning output of a FLUX Redux invocation. */ - FaceMaskInvocation: { + FluxReduxOutput: { /** - * @description Optional metadata to be saved with the image - * @default null + * Conditioning + * @description FLUX Redux conditioning tensor */ - metadata?: components["schemas"]["MetadataField"] | null; + redux_cond: components["schemas"]["FluxReduxConditioningField"]; + /** + * type + * @default flux_redux_output + * @constant + */ + type: "flux_redux_output"; + }; + /** + * Prompt - FLUX + * @description Encodes and preps a prompt for a flux image. + */ + FluxTextEncoderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11478,84 +10419,51 @@ export type components = { */ use_cache?: boolean; /** - * @description Image to face detect + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * Face Ids - * @description Comma-separated list of face ids to mask eg '0,2,7'. Numbered from 0. Leave empty to mask all. Find face IDs with FaceIdentifier node. - * @default - */ - face_ids?: string; - /** - * Minimum Confidence - * @description Minimum confidence for face detection (lower if detection is failing) - * @default 0.5 - */ - minimum_confidence?: number; + clip?: components["schemas"]["CLIPField"] | null; /** - * X Offset - * @description Offset for the X-axis of the face mask - * @default 0 + * T5Encoder + * @description T5 tokenizer and text encoder + * @default null */ - x_offset?: number; + t5_encoder?: components["schemas"]["T5EncoderField"] | null; /** - * Y Offset - * @description Offset for the Y-axis of the face mask - * @default 0 + * T5 Max Seq Len + * @description Max sequence length for the T5 encoder. Expected to be 256 for FLUX schnell models and 512 for FLUX dev models. + * @default null */ - y_offset?: number; + t5_max_seq_len?: (256 | 512) | null; /** - * Chunk - * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. - * @default false + * Prompt + * @description Text prompt to encode. + * @default null */ - chunk?: boolean; + prompt?: string | null; /** - * Invert Mask - * @description Toggle to invert the mask - * @default false + * @description A mask defining the region that this conditioning prompt applies to. + * @default null */ - invert_mask?: boolean; + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default face_mask_detection + * @default flux_text_encoder * @constant */ - type: "face_mask_detection"; + type: "flux_text_encoder"; }; /** - * FaceMaskOutput - * @description Base class for FaceMask output + * Latents to Image - FLUX + * @description Generates an image from latents. */ - FaceMaskOutput: { - /** @description The output image */ - image: components["schemas"]["ImageField"]; - /** - * Width - * @description The width of the image in pixels - */ - width: number; - /** - * Height - * @description The height of the image in pixels - */ - height: number; + FluxVaeDecodeInvocation: { /** - * type - * @default face_mask_output - * @constant + * @description The board to save the image to + * @default null */ - type: "face_mask_output"; - /** @description The output mask */ - mask: components["schemas"]["ImageField"]; - }; - /** - * FaceOff - * @description Bound, extract, and mask a face from an image using MediaPipe detection - */ - FaceOffInvocation: { + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -11579,122 +10487,27 @@ export type components = { */ use_cache?: boolean; /** - * @description Image for face detection + * @description Latents tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * Face Id - * @description The face ID to process, numbered from 0. Multiple faces not supported. Find a face's ID with FaceIdentifier node. - * @default 0 + * @description VAE + * @default null */ - face_id?: number; + vae?: components["schemas"]["VAEField"] | null; /** - * Minimum Confidence - * @description Minimum confidence for face detection (lower if detection is failing) - * @default 0.5 + * type + * @default flux_vae_decode + * @constant */ - minimum_confidence?: number; - /** - * X Offset - * @description X-axis offset of the mask - * @default 0 - */ - x_offset?: number; - /** - * Y Offset - * @description Y-axis offset of the mask - * @default 0 - */ - y_offset?: number; - /** - * Padding - * @description All-axis padding around the mask in pixels - * @default 0 - */ - padding?: number; - /** - * Chunk - * @description Whether to bypass full image face detection and default to image chunking. Chunking will occur if no faces are found in the full image. - * @default false - */ - chunk?: boolean; - /** - * type - * @default face_off - * @constant - */ - type: "face_off"; - }; - /** - * FaceOffOutput - * @description Base class for FaceOff Output - */ - FaceOffOutput: { - /** @description The output image */ - image: components["schemas"]["ImageField"]; - /** - * Width - * @description The width of the image in pixels - */ - width: number; - /** - * Height - * @description The height of the image in pixels - */ - height: number; - /** - * type - * @default face_off_output - * @constant - */ - type: "face_off_output"; - /** @description The output mask */ - mask: components["schemas"]["ImageField"]; - /** - * X - * @description The x coordinate of the bounding box's left side - */ - x: number; - /** - * Y - * @description The y coordinate of the bounding box's top side - */ - y: number; + type: "flux_vae_decode"; }; /** - * FieldKind - * @description The kind of field. - * - `Input`: An input field on a node. - * - `Output`: An output field on a node. - * - `Internal`: A field which is treated as an input, but cannot be used in node definitions. Metadata is - * one example. It is provided to nodes via the WithMetadata class, and we want to reserve the field name - * "metadata" for this on all nodes. `FieldKind` is used to short-circuit the field name validation logic, - * allowing "metadata" for that field. - * - `NodeAttribute`: The field is a node attribute. These are fields which are not inputs or outputs, - * but which are used to store information about the node. For example, the `id` and `type` fields are node - * attributes. - * - * The presence of this in `json_schema_extra["field_kind"]` is used when initializing node schemas on app - * startup, and when generating the OpenAPI schema for the workflow editor. - * @enum {string} - */ - FieldKind: "input" | "output" | "internal" | "node_attribute"; - /** - * FilmGrain - * @description Adds film grain to an image + * Image to Latents - FLUX + * @description Encodes an image into latents. */ - FilmGrainInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + FluxVaeEncodeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11713,102 +10526,78 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to add film grain to + * @description The image to encode. * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Amount 1 - * @description Amount of the first noise layer - * @default 100 - */ - amount_1?: number; - /** - * Amount 2 - * @description Amount of the second noise layer - * @default 50 - */ - amount_2?: number; - /** - * Seed 1 - * @description The first seed to use (omit for random) - * @default null - */ - seed_1?: number | null; - /** - * Seed 2 - * @description The second seed to use (omit for random) + * @description VAE * @default null */ - seed_2?: number | null; - /** - * Blur 1 - * @description The strength of the first noise blur - * @default 0.5 - */ - blur_1?: number; - /** - * Blur 2 - * @description The strength of the second noise blur - * @default 0.5 - */ - blur_2?: number; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default film_grain + * @default flux_vae_encode * @constant */ - type: "film_grain"; + type: "flux_vae_encode"; }; /** - * Flatten Histogram (Grayscale) - * @description Scales the values of an L-mode image by scaling them to the full range 0..255 in equal proportions + * FluxVariantType + * @description FLUX.1 model variants. + * @enum {string} */ - FlattenHistogramMono: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + FluxVariantType: "schnell" | "dev" | "dev_fill"; + /** FoundModel */ + FoundModel: { /** - * @description Optional metadata to be saved with the image - * @default null + * Path + * @description Path to the model */ - metadata?: components["schemas"]["MetadataField"] | null; + path: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Is Installed + * @description Whether or not the model is already installed */ - id: string; + is_installed: boolean; + }; + /** + * FreeUConfig + * @description Configuration for the FreeU hyperparameters. + * - https://huggingface.co/docs/diffusers/main/en/using-diffusers/freeu + * - https://github.com/ChenyangSi/FreeU + */ + FreeUConfig: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * S1 + * @description Scaling factor for stage 1 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. */ - is_intermediate?: boolean; + s1: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * S2 + * @description Scaling factor for stage 2 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. */ - use_cache?: boolean; + s2: number; /** - * @description Single-channel image for which to flatten the histogram - * @default null + * B1 + * @description Scaling factor for stage 1 to amplify the contributions of backbone features. */ - image?: components["schemas"]["ImageField"] | null; + b1: number; /** - * type - * @default flatten_histogram_mono - * @constant + * B2 + * @description Scaling factor for stage 2 to amplify the contributions of backbone features. */ - type: "flatten_histogram_mono"; + b2: number; }; /** - * Float Batch - * @description Create a batched generation, where the workflow is executed once for each float in the batch. + * Apply FreeU - SD1.5, SDXL + * @description Applies FreeU to the UNet. Suggested values (b1/b2/s1/s2): + * + * SD1.5: 1.2/1.4/0.9/0.2, + * SD2: 1.1/1.2/0.9/0.2, + * SDXL: 1.1/1.2/0.6/0.4, */ - FloatBatchInvocation: { + FreeUInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11827,30 +10616,58 @@ export type components = { */ use_cache?: boolean; /** - * Batch Group - * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. - * @default None - * @enum {string} + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; + unet?: components["schemas"]["UNetField"] | null; /** - * Floats - * @description The floats to batch over - * @default null + * B1 + * @description Scaling factor for stage 1 to amplify the contributions of backbone features. + * @default 1.2 */ - floats?: number[] | null; + b1?: number; + /** + * B2 + * @description Scaling factor for stage 2 to amplify the contributions of backbone features. + * @default 1.4 + */ + b2?: number; + /** + * S1 + * @description Scaling factor for stage 1 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. + * @default 0.9 + */ + s1?: number; + /** + * S2 + * @description Scaling factor for stage 2 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. + * @default 0.2 + */ + s2?: number; /** * type - * @default float_batch + * @default freeu * @constant */ - type: "float_batch"; + type: "freeu"; + }; + /** + * GeneratePasswordResponse + * @description Response containing a generated password. + */ + GeneratePasswordResponse: { + /** + * Password + * @description Generated strong password + */ + password: string; }; /** - * Float Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option + * Get Image Mask Bounding Box + * @description Gets the bounding box of the given mask image. */ - FloatCollectionIndexInvocation: { + GetMaskBoundingBoxInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -11865,132 +10682,147 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Random - * @description Random Index? - * @default true + * @description The mask to crop. + * @default null */ - random?: boolean; + mask?: components["schemas"]["ImageField"] | null; /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) + * Margin + * @description Margin to add to the bounding box. * @default 0 */ - index?: number; + margin?: number; /** - * Collection - * @description float collection - * @default null + * @description Color of the mask in the image. + * @default { + * "r": 255, + * "g": 255, + * "b": 255, + * "a": 255 + * } */ - collection?: number[] | null; + mask_color?: components["schemas"]["ColorField"]; /** * type - * @default float_collection_index + * @default get_image_mask_bounding_box * @constant */ - type: "float_collection_index"; + type: "get_image_mask_bounding_box"; + }; + /** GlmEncoderField */ + GlmEncoderField: { + /** @description Info to load tokenizer submodel */ + tokenizer: components["schemas"]["ModelIdentifierField"]; + /** @description Info to load text_encoder submodel */ + text_encoder: components["schemas"]["ModelIdentifierField"]; }; /** - * Float Collection Primitive - * @description A collection of float primitive values + * GradientMaskOutput + * @description Outputs a denoise mask and an image representing the total gradient of the mask. */ - FloatCollectionInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + GradientMaskOutput: { + /** @description Mask for denoise model run. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. */ + denoise_mask: components["schemas"]["DenoiseMaskField"]; + /** @description Image representing the total gradient area of the mask. For paste-back purposes. */ + expanded_mask_area: components["schemas"]["ImageField"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * type + * @default gradient_mask_output + * @constant */ - is_intermediate?: boolean; + type: "gradient_mask_output"; + }; + /** Graph */ + Graph: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Id + * @description The id of this graph */ - use_cache?: boolean; + id?: string; /** - * Collection - * @description The collection of float values - * @default [] + * Nodes + * @description The nodes in this graph */ - collection?: number[]; + nodes?: { + [key: string]: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; + }; /** - * type - * @default float_collection - * @constant + * Edges + * @description The connections between nodes and their fields in this graph */ - type: "float_collection"; + edges?: components["schemas"]["Edge"][]; }; /** - * Float Collection Primitive linked - * @description A collection of float primitive values + * GraphExecutionState + * @description Tracks the state of a graph execution */ - FloatCollectionLinkedInvocation: { + GraphExecutionState: { /** * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description The id of the execution state */ id: string; + /** @description The graph being executed */ + graph: components["schemas"]["Graph"]; + /** @description The expanded graph of activated and executed nodes */ + execution_graph: components["schemas"]["Graph"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Executed + * @description The set of node ids that have been executed */ - is_intermediate?: boolean; + executed: string[]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Executed History + * @description The list of node ids that have been executed, in order of execution */ - use_cache?: boolean; + executed_history: string[]; /** - * Collection - * @description The collection of float values - * @default [] + * Results + * @description The results of node executions */ - collection?: number[]; + results: { + [key: string]: components["schemas"]["BooleanCollectionOutput"] | components["schemas"]["BooleanOutput"] | components["schemas"]["BoundingBoxCollectionOutput"] | components["schemas"]["BoundingBoxOutput"] | components["schemas"]["CLIPOutput"] | components["schemas"]["CLIPSkipInvocationOutput"] | components["schemas"]["CalculateImageTilesOutput"] | components["schemas"]["CogView4ConditioningOutput"] | components["schemas"]["CogView4ModelLoaderOutput"] | components["schemas"]["CollectInvocationOutput"] | components["schemas"]["ColorCollectionOutput"] | components["schemas"]["ColorOutput"] | components["schemas"]["ConditioningCollectionOutput"] | components["schemas"]["ConditioningOutput"] | components["schemas"]["ControlOutput"] | components["schemas"]["DenoiseMaskOutput"] | components["schemas"]["FaceMaskOutput"] | components["schemas"]["FaceOffOutput"] | components["schemas"]["FloatCollectionOutput"] | components["schemas"]["FloatGeneratorOutput"] | components["schemas"]["FloatOutput"] | components["schemas"]["Flux2KleinLoRALoaderOutput"] | components["schemas"]["Flux2KleinModelLoaderOutput"] | components["schemas"]["FluxConditioningCollectionOutput"] | components["schemas"]["FluxConditioningOutput"] | components["schemas"]["FluxControlLoRALoaderOutput"] | components["schemas"]["FluxControlNetOutput"] | components["schemas"]["FluxFillOutput"] | components["schemas"]["FluxKontextOutput"] | components["schemas"]["FluxLoRALoaderOutput"] | components["schemas"]["FluxModelLoaderOutput"] | components["schemas"]["FluxReduxOutput"] | components["schemas"]["GradientMaskOutput"] | components["schemas"]["IPAdapterOutput"] | components["schemas"]["IdealSizeOutput"] | components["schemas"]["ImageCollectionOutput"] | components["schemas"]["ImageGeneratorOutput"] | components["schemas"]["ImageOutput"] | components["schemas"]["ImagePanelCoordinateOutput"] | components["schemas"]["IntegerCollectionOutput"] | components["schemas"]["IntegerGeneratorOutput"] | components["schemas"]["IntegerOutput"] | components["schemas"]["IterateInvocationOutput"] | components["schemas"]["LatentsCollectionOutput"] | components["schemas"]["LatentsMetaOutput"] | components["schemas"]["LatentsOutput"] | components["schemas"]["LoRALoaderOutput"] | components["schemas"]["LoRASelectorOutput"] | components["schemas"]["MDControlListOutput"] | components["schemas"]["MDIPAdapterListOutput"] | components["schemas"]["MDT2IAdapterListOutput"] | components["schemas"]["MaskOutput"] | components["schemas"]["MetadataItemOutput"] | components["schemas"]["MetadataOutput"] | components["schemas"]["MetadataToLorasCollectionOutput"] | components["schemas"]["MetadataToModelOutput"] | components["schemas"]["MetadataToSDXLModelOutput"] | components["schemas"]["ModelIdentifierOutput"] | components["schemas"]["ModelLoaderOutput"] | components["schemas"]["NoiseOutput"] | components["schemas"]["PBRMapsOutput"] | components["schemas"]["PairTileImageOutput"] | components["schemas"]["PromptTemplateOutput"] | components["schemas"]["SD3ConditioningOutput"] | components["schemas"]["SDXLLoRALoaderOutput"] | components["schemas"]["SDXLModelLoaderOutput"] | components["schemas"]["SDXLRefinerModelLoaderOutput"] | components["schemas"]["SchedulerOutput"] | components["schemas"]["Sd3ModelLoaderOutput"] | components["schemas"]["SeamlessModeOutput"] | components["schemas"]["String2Output"] | components["schemas"]["StringCollectionOutput"] | components["schemas"]["StringGeneratorOutput"] | components["schemas"]["StringOutput"] | components["schemas"]["StringPosNegOutput"] | components["schemas"]["T2IAdapterOutput"] | components["schemas"]["TileToPropertiesOutput"] | components["schemas"]["UNetOutput"] | components["schemas"]["VAEOutput"] | components["schemas"]["ZImageConditioningOutput"] | components["schemas"]["ZImageControlOutput"] | components["schemas"]["ZImageLoRALoaderOutput"] | components["schemas"]["ZImageModelLoaderOutput"]; + }; /** - * type - * @default float_collection_linked - * @constant + * Errors + * @description Errors raised when executing nodes */ - type: "float_collection_linked"; + errors: { + [key: string]: string; + }; /** - * Value - * @description The float value - * @default 0 + * Prepared Source Mapping + * @description The map of prepared nodes to original graph nodes */ - value?: number; - }; - /** - * FloatCollectionOutput - * @description Base class for nodes that output a collection of floats - */ - FloatCollectionOutput: { + prepared_source_mapping: { + [key: string]: string; + }; /** - * Collection - * @description The float collection + * Source Prepared Mapping + * @description The map of original graph nodes to prepared nodes */ - collection: number[]; + source_prepared_mapping: { + [key: string]: string[]; + }; + /** Ready Order */ + ready_order?: string[]; /** - * type - * @default float_collection_output - * @constant + * Indegree + * @description Remaining unmet input count for exec nodes */ - type: "float_collection_output"; + indegree?: { + [key: string]: number; + }; }; /** - * Float Collection Toggle - * @description Allows boolean selection between two separate float collection inputs + * Grounding DINO (Text Prompt Object Detection) + * @description Runs a Grounding DINO model. Performs zero-shot bounding-box object detection from a text prompt. */ - FloatCollectionToggleInvocation: { + GroundingDinoInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12009,35 +10841,50 @@ export type components = { */ use_cache?: boolean; /** - * Use Second - * @description Use 2nd Input - * @default false + * Model + * @description The Grounding DINO model to use. + * @default null */ - use_second?: boolean; + model?: ("grounding-dino-tiny" | "grounding-dino-base") | null; /** - * Col1 - * @description First Float Collection Input + * Prompt + * @description The prompt describing the object to segment. * @default null */ - col1?: number[] | null; + prompt?: string | null; /** - * Col2 - * @description Second Float Collection Input + * @description The image to segment. * @default null */ - col2?: number[] | null; + image?: components["schemas"]["ImageField"] | null; + /** + * Detection Threshold + * @description The detection threshold for the Grounding DINO model. All detected bounding boxes with scores above this threshold will be returned. + * @default 0.3 + */ + detection_threshold?: number; /** * type - * @default float_collection_toggle + * @default grounding_dino * @constant */ - type: "float_collection_toggle"; + type: "grounding_dino"; }; /** - * Float Generator - * @description Generated a range of floats for use in a batched generation + * HED Edge Detection + * @description Geneartes an edge map using the HED (softedge) model. */ - FloatGenerator: { + HEDEdgeDetectionInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12056,41 +10903,62 @@ export type components = { */ use_cache?: boolean; /** - * Generator Type - * @description The float generator. + * @description The image to process + * @default null */ - generator: components["schemas"]["FloatGeneratorField"]; + image?: components["schemas"]["ImageField"] | null; + /** + * Scribble + * @description Whether or not to use scribble mode + * @default false + */ + scribble?: boolean; /** * type - * @default float_generator + * @default hed_edge_detection * @constant */ - type: "float_generator"; + type: "hed_edge_detection"; }; - /** FloatGeneratorField */ - FloatGeneratorField: Record; /** - * FloatGeneratorOutput - * @description Base class for nodes that output a collection of floats + * HFModelSource + * @description A HuggingFace repo_id with optional variant, sub-folder(s) and access token. + * Note that the variant option, if not provided to the constructor, will default to fp16, which is + * what people (almost) always want. + * + * The subfolder can be a single path or multiple paths joined by '+' (e.g., "text_encoder+tokenizer"). + * When multiple subfolders are specified, all of them will be downloaded and combined into the model directory. */ - FloatGeneratorOutput: { - /** - * Floats - * @description The generated floats - */ - floats: number[]; + HFModelSource: { + /** Repo Id */ + repo_id: string; + /** @default fp16 */ + variant?: components["schemas"]["ModelRepoVariant"] | null; + /** Subfolder */ + subfolder?: string | null; + /** Access Token */ + access_token?: string | null; /** - * type - * @default float_generator_output - * @constant + * @description discriminator enum property added by openapi-typescript + * @enum {string} */ - type: "float_generator_output"; + type: "hf"; }; /** - * Float Primitive - * @description A float primitive value + * HFTokenStatus + * @enum {string} */ - FloatInvocation: { + HFTokenStatus: "valid" | "invalid" | "unknown"; + /** HTTPValidationError */ + HTTPValidationError: { + /** Detail */ + detail?: components["schemas"]["ValidationError"][]; + }; + /** + * Heuristic Resize + * @description Resize an image using a heuristic method. Preserves edge maps. + */ + HeuristicResizeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12109,135 +10977,136 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The float value - * @default 0 + * @description The image to resize + * @default null */ - value?: number; + image?: components["schemas"]["ImageField"] | null; + /** + * Width + * @description The width to resize to (px) + * @default 512 + */ + width?: number; + /** + * Height + * @description The height to resize to (px) + * @default 512 + */ + height?: number; /** * type - * @default float + * @default heuristic_resize * @constant */ - type: "float"; + type: "heuristic_resize"; }; /** - * Float Range - * @description Creates a range + * HuggingFaceMetadata + * @description Extended metadata fields provided by HuggingFace. */ - FloatLinearRangeInvocation: { + HuggingFaceMetadata: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Name + * @description model's name */ - id: string; + name: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Files + * @description model files and their sizes */ - is_intermediate?: boolean; + files?: components["schemas"]["RemoteModelFile"][]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * @description discriminator enum property added by openapi-typescript + * @enum {string} */ - use_cache?: boolean; + type: "huggingface"; /** - * Start - * @description The first value of the range - * @default 5 + * Id + * @description The HF model id */ - start?: number; + id: string; /** - * Stop - * @description The last value of the range - * @default 10 + * Api Response + * @description Response from the HF API as stringified JSON */ - stop?: number; + api_response?: string | null; /** - * Steps - * @description number of values to interpolate over (including start and stop) - * @default 30 + * Is Diffusers + * @description Whether the metadata is for a Diffusers format model + * @default false */ - steps?: number; + is_diffusers?: boolean; /** - * type - * @default float_range - * @constant + * Ckpt Urls + * @description URLs for all checkpoint format models in the metadata */ - type: "float_range"; + ckpt_urls?: string[] | null; }; - /** - * Float Math - * @description Performs floating point math. - */ - FloatMathInvocation: { + /** HuggingFaceModels */ + HuggingFaceModels: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Urls + * @description URLs for all checkpoint format models in the metadata */ - id: string; + urls: string[] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Is Diffusers + * @description Whether the metadata is for a Diffusers format model */ - is_intermediate?: boolean; + is_diffusers: boolean; + }; + /** IPAdapterField */ + IPAdapterField: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Image + * @description The IP-Adapter image prompt(s). */ - use_cache?: boolean; + image: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][]; + /** @description The IP-Adapter model to use. */ + ip_adapter_model: components["schemas"]["ModelIdentifierField"]; + /** @description The name of the CLIP image encoder model. */ + image_encoder_model: components["schemas"]["ModelIdentifierField"]; /** - * Operation - * @description The operation to perform - * @default ADD - * @enum {string} + * Weight + * @description The weight given to the IP-Adapter. + * @default 1 */ - operation?: "ADD" | "SUB" | "MUL" | "DIV" | "EXP" | "ABS" | "SQRT" | "MIN" | "MAX"; + weight?: number | number[]; /** - * A - * @description The first number - * @default 1 + * Target Blocks + * @description The IP Adapter blocks to apply + * @default [] */ - a?: number; + target_blocks?: string[]; /** - * B - * @description The second number - * @default 1 + * Method + * @description Weight apply method + * @default full */ - b?: number; + method?: string; /** - * type - * @default float_math - * @constant + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) + * @default 0 */ - type: "float_math"; - }; - /** - * FloatOutput - * @description Base class for nodes that output a single float - */ - FloatOutput: { + begin_step_percent?: number; /** - * Value - * @description The output float + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) + * @default 1 */ - value: number; + end_step_percent?: number; /** - * type - * @default float_output - * @constant + * @description The bool mask associated with this IP-Adapter. Excluded regions should be set to False, included regions should be set to True. + * @default null */ - type: "float_output"; + mask?: components["schemas"]["TensorField"] | null; }; /** - * Float To Integer - * @description Rounds a float number to (a multiple of) an integer. + * IP-Adapter - SD1.5, SDXL + * @description Collects IP-Adapter info to pass to other nodes. */ - FloatToIntegerInvocation: { + IPAdapterInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12256,619 +11125,641 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The value to round - * @default 0 + * Image + * @description The IP-Adapter image prompt(s). + * @default null */ - value?: number; + image?: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][] | null; /** - * Multiple of - * @description The multiple to round to + * IP-Adapter Model + * @description The IP-Adapter model. + * @default null + */ + ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Clip Vision Model + * @description CLIP Vision model to use. Overrides model settings. Mandatory for checkpoint models. + * @default ViT-H + * @enum {string} + */ + clip_vision_model?: "ViT-H" | "ViT-G" | "ViT-L"; + /** + * Weight + * @description The weight given to the IP-Adapter * @default 1 */ - multiple?: number; + weight?: number | number[]; /** * Method - * @description The method to use for rounding - * @default Nearest + * @description The method to apply the IP-Adapter + * @default full * @enum {string} */ - method?: "Nearest" | "Floor" | "Ceiling" | "Truncate"; + method?: "full" | "style" | "composition" | "style_strong" | "style_precise"; + /** + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) + * @default 0 + */ + begin_step_percent?: number; + /** + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) + * @default 1 + */ + end_step_percent?: number; + /** + * @description A mask defining the region that this IP-Adapter applies to. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default float_to_int + * @default ip_adapter * @constant */ - type: "float_to_int"; + type: "ip_adapter"; }; /** - * Float Toggle - * @description Allows boolean selection between two separate float inputs + * IPAdapterMetadataField + * @description IP Adapter Field, minus the CLIP Vision Encoder model */ - FloatToggleInvocation: { + IPAdapterMetadataField: { + /** @description The IP-Adapter image prompt. */ + image: components["schemas"]["ImageField"]; + /** @description The IP-Adapter model. */ + ip_adapter_model: components["schemas"]["ModelIdentifierField"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Clip Vision Model + * @description The CLIP Vision model + * @enum {string} */ - id: string; + clip_vision_model: "ViT-L" | "ViT-H" | "ViT-G"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Method + * @description Method to apply IP Weights with + * @enum {string} */ - is_intermediate?: boolean; + method: "full" | "style" | "composition" | "style_strong" | "style_precise"; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Weight + * @description The weight given to the IP-Adapter */ - use_cache?: boolean; + weight: number | number[]; /** - * Use Second - * @description Use 2nd Input - * @default false + * Begin Step Percent + * @description When the IP-Adapter is first applied (% of total steps) */ - use_second?: boolean; + begin_step_percent: number; /** - * Float1 - * @description First Float Input - * @default 0 + * End Step Percent + * @description When the IP-Adapter is last applied (% of total steps) */ - float1?: number; + end_step_percent: number; + }; + /** IPAdapterOutput */ + IPAdapterOutput: { /** - * Float2 - * @description Second Float Input - * @default 0 + * IP-Adapter + * @description IP-Adapter to apply */ - float2?: number; + ip_adapter: components["schemas"]["IPAdapterField"]; /** * type - * @default float_toggle + * @default ip_adapter_output * @constant */ - type: "float_toggle"; + type: "ip_adapter_output"; }; /** - * Floats To Strings - * @description Converts a float or collections of floats to a collection of strings + * IPAdapterRecallParameter + * @description IP Adapter configuration for recall */ - FloatsToStringsInvocation: { + IPAdapterRecallParameter: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Model Name + * @description The name of the IP Adapter model */ - id: string; + model_name: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Image Name + * @description The filename of the reference image in outputs/images */ - use_cache?: boolean; + image_name?: string | null; /** - * Floats - * @description float or collection of floats - * @default [] + * Weight + * @description The weight for the IP Adapter + * @default 1 */ - floats?: number | number[]; + weight?: number; /** - * type - * @default floats_to_strings - * @constant + * Begin Step Percent + * @description When the IP Adapter is first applied (% of total steps) */ - type: "floats_to_strings"; - }; - /** - * FLUX2 Denoise - * @description Run denoising process with a FLUX.2 Klein transformer model. - * - * This node is designed for FLUX.2 Klein models which use Qwen3 as the text encoder. - * It does not support ControlNet, IP-Adapters, or regional prompting. - */ - Flux2DenoiseInvocation: { + begin_step_percent?: number | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * End Step Percent + * @description When the IP Adapter is last applied (% of total steps) */ - id: string; + end_step_percent?: number | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Method + * @description The IP Adapter method */ - is_intermediate?: boolean; + method?: ("full" | "style" | "composition") | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Image Influence + * @description FLUX Redux image influence (if model is flux_redux) */ - use_cache?: boolean; + image_influence?: ("lowest" | "low" | "medium" | "high" | "highest") | null; + }; + /** IPAdapter_Checkpoint_FLUX_Config */ + IPAdapter_Checkpoint_FLUX_Config: { /** - * @description Latents tensor - * @default null + * Key + * @description A unique key for this model. */ - latents?: components["schemas"]["LatentsField"] | null; + key: string; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. - * @default null + * Hash + * @description The hash of the model file(s). */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + hash: string; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - denoising_start?: number; + path: string; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * File Size + * @description The size of the model in bytes. */ - denoising_end?: number; + file_size: number; /** - * Add Noise - * @description Add noise based on denoising start. - * @default true + * Name + * @description Name of the model. */ - add_noise?: boolean; + name: string; /** - * Transformer - * @description Flux model (Transformer) to load - * @default null + * Description + * @description Model description */ - transformer?: components["schemas"]["TransformerField"] | null; + description: string | null; /** - * @description Positive conditioning tensor - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | null; + source_api_response: string | null; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 1 + * Cover Image + * @description Url for image to preview model */ - cfg_scale?: number; + cover_image: string | null; /** - * Width - * @description Width of the generated image. - * @default 1024 + * Type + * @default ip_adapter + * @constant */ - width?: number; + type: "ip_adapter"; /** - * Height - * @description Height of the generated image. - * @default 1024 + * Format + * @default checkpoint + * @constant */ - height?: number; + format: "checkpoint"; /** - * Num Steps - * @description Number of diffusion steps. Use 4 for distilled models, 28+ for base models. - * @default 4 + * Base + * @default flux + * @constant */ - num_steps?: number; + base: "flux"; + }; + /** IPAdapter_Checkpoint_SD1_Config */ + IPAdapter_Checkpoint_SD1_Config: { /** - * Scheduler - * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. - * @default euler - * @enum {string} + * Key + * @description A unique key for this model. */ - scheduler?: "euler" | "heun" | "lcm"; + key: string; /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * Hash + * @description The hash of the model file(s). */ - seed?: number; + hash: string; /** - * @description FLUX.2 VAE model (required for BN statistics). - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - vae?: components["schemas"]["VAEField"] | null; + path: string; /** - * Reference Images - * @description FLUX Kontext conditioning (reference images for multi-reference image editing). - * @default null + * File Size + * @description The size of the model in bytes. */ - kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; + file_size: number; /** - * type - * @default flux2_denoise - * @constant + * Name + * @description Name of the model. */ - type: "flux2_denoise"; - }; - /** - * Apply LoRA Collection - Flux2 Klein - * @description Applies a collection of LoRAs to a FLUX.2 Klein transformer and/or Qwen3 text encoder. - */ - Flux2KleinLoRACollectionLoader: { + name: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Description + * @description Model description */ - id: string; + description: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - is_intermediate?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - use_cache?: boolean; + source_api_response: string | null; /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. - * @default null + * Cover Image + * @description Url for image to preview model */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; + cover_image: string | null; /** - * Transformer - * @description Transformer - * @default null + * Type + * @default ip_adapter + * @constant */ - transformer?: components["schemas"]["TransformerField"] | null; + type: "ip_adapter"; /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder - * @default null + * Format + * @default checkpoint + * @constant */ - qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; + format: "checkpoint"; /** - * type - * @default flux2_klein_lora_collection_loader + * Base + * @default sd-1 * @constant */ - type: "flux2_klein_lora_collection_loader"; + base: "sd-1"; }; - /** - * Apply LoRA - Flux2 Klein - * @description Apply a LoRA model to a FLUX.2 Klein transformer and/or Qwen3 text encoder. - */ - Flux2KleinLoRALoaderInvocation: { + /** IPAdapter_Checkpoint_SD2_Config */ + IPAdapter_Checkpoint_SD2_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRA - * @description LoRA model to load - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - lora?: components["schemas"]["ModelIdentifierField"] | null; + path: string; /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 + * File Size + * @description The size of the model in bytes. */ - weight?: number; + file_size: number; /** - * Transformer - * @description Transformer - * @default null + * Name + * @description Name of the model. */ - transformer?: components["schemas"]["TransformerField"] | null; + name: string; /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder - * @default null + * Description + * @description Model description */ - qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; + description: string | null; /** - * type - * @default flux2_klein_lora_loader - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "flux2_klein_lora_loader"; - }; - /** - * Flux2KleinLoRALoaderOutput - * @description FLUX.2 Klein LoRA Loader Output - */ - Flux2KleinLoRALoaderOutput: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Transformer - * @description Transformer - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - transformer: components["schemas"]["TransformerField"] | null; + source_api_response: string | null; /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder - * @default null + * Cover Image + * @description Url for image to preview model */ - qwen3_encoder: components["schemas"]["Qwen3EncoderField"] | null; + cover_image: string | null; /** - * type - * @default flux2_klein_lora_loader_output + * Type + * @default ip_adapter * @constant */ - type: "flux2_klein_lora_loader_output"; - }; - /** - * Main Model - Flux2 Klein - * @description Loads a Flux2 Klein model, outputting its submodels. - * - * Flux2 Klein uses Qwen3 as the text encoder instead of CLIP+T5. - * It uses a 32-channel VAE (AutoencoderKLFlux2) instead of the 16-channel FLUX.1 VAE. - * - * When using a Diffusers format model, both VAE and Qwen3 encoder are extracted - * automatically from the main model. You can override with standalone models: - * - Transformer: Always from Flux2 Klein main model - * - VAE: From main model (Diffusers) or standalone VAE - * - Qwen3 Encoder: From main model (Diffusers) or standalone Qwen3 model - */ - Flux2KleinModelLoaderInvocation: { + type: "ip_adapter"; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Format + * @default checkpoint + * @constant */ - id: string; + format: "checkpoint"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Base + * @default sd-2 + * @constant */ - is_intermediate?: boolean; + base: "sd-2"; + }; + /** IPAdapter_Checkpoint_SDXL_Config */ + IPAdapter_Checkpoint_SDXL_Config: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Key + * @description A unique key for this model. */ - use_cache?: boolean; + key: string; /** - * Transformer - * @description Flux model (Transformer) to load + * Hash + * @description The hash of the model file(s). */ - model: components["schemas"]["ModelIdentifierField"]; + hash: string; /** - * VAE - * @description Standalone VAE model. Flux2 Klein uses the same VAE as FLUX (16-channel). If not provided, VAE will be loaded from the Qwen3 Source model. - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + path: string; /** - * Qwen3 Encoder - * @description Standalone Qwen3 Encoder model. If not provided, encoder will be loaded from the Qwen3 Source model. - * @default null + * File Size + * @description The size of the model in bytes. */ - qwen3_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; + file_size: number; /** - * Qwen3 Source (Diffusers) - * @description Diffusers Flux2 Klein model to extract VAE and/or Qwen3 encoder from. Use this if you don't have separate VAE/Qwen3 models. Ignored if both VAE and Qwen3 Encoder are provided separately. - * @default null + * Name + * @description Name of the model. */ - qwen3_source_model?: components["schemas"]["ModelIdentifierField"] | null; + name: string; /** - * Max Seq Length - * @description Max sequence length for the Qwen3 encoder. - * @default 512 - * @enum {integer} + * Description + * @description Model description */ - max_seq_len?: 256 | 512; + description: string | null; /** - * type - * @default flux2_klein_model_loader - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "flux2_klein_model_loader"; - }; - /** - * Flux2KleinModelLoaderOutput - * @description Flux2 Klein model loader output. - */ - Flux2KleinModelLoaderOutput: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Transformer - * @description Transformer + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - transformer: components["schemas"]["TransformerField"]; + source_api_response: string | null; /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder + * Cover Image + * @description Url for image to preview model */ - qwen3_encoder: components["schemas"]["Qwen3EncoderField"]; + cover_image: string | null; /** - * VAE - * @description VAE + * Type + * @default ip_adapter + * @constant */ - vae: components["schemas"]["VAEField"]; + type: "ip_adapter"; /** - * Max Seq Length - * @description The max sequence length for the Qwen3 encoder. - * @enum {integer} + * Format + * @default checkpoint + * @constant */ - max_seq_len: 256 | 512; + format: "checkpoint"; /** - * type - * @default flux2_klein_model_loader_output + * Base + * @default sdxl * @constant */ - type: "flux2_klein_model_loader_output"; + base: "sdxl"; }; - /** - * Prompt - Flux2 Klein - * @description Encodes and preps a prompt for Flux2 Klein image generation. - * - * Flux2 Klein uses Qwen3 as the text encoder, extracting hidden states from - * layers (9, 18, 27) and stacking them for richer text representations. - * This matches the diffusers Flux2KleinPipeline implementation exactly. - */ - Flux2KleinTextEncoderInvocation: { + /** IPAdapter_InvokeAI_SD1_Config */ + IPAdapter_InvokeAI_SD1_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Prompt - * @description Text prompt to encode. - * @default null + * File Size + * @description The size of the model in bytes. */ - prompt?: string | null; + file_size: number; /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder - * @default null + * Name + * @description Name of the model. */ - qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; + name: string; /** - * Max Seq Len - * @description Max sequence length for the Qwen3 encoder. - * @default 512 - * @enum {integer} + * Description + * @description Model description */ - max_seq_len?: 256 | 512; + description: string | null; /** - * @description A mask defining the region that this conditioning prompt applies to. - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - mask?: components["schemas"]["TensorField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * type - * @default flux2_klein_text_encoder - * @constant - */ - type: "flux2_klein_text_encoder"; + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter + * @constant + */ + type: "ip_adapter"; + /** + * Format + * @default invokeai + * @constant + */ + format: "invokeai"; + /** Image Encoder Model Id */ + image_encoder_model_id: string; + /** + * Base + * @default sd-1 + * @constant + */ + base: "sd-1"; }; - /** - * Latents to Image - FLUX2 - * @description Generates an image from latents using FLUX.2 Klein's 32-channel VAE. - */ - Flux2VaeDecodeInvocation: { + /** IPAdapter_InvokeAI_SD2_Config */ + IPAdapter_InvokeAI_SD2_Config: { /** - * @description The board to save the image to - * @default null + * Key + * @description A unique key for this model. */ - board?: components["schemas"]["BoardField"] | null; + key: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Hash + * @description The hash of the model file(s). */ - metadata?: components["schemas"]["MetadataField"] | null; + hash: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - id: string; + path: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * File Size + * @description The size of the model in bytes. */ - is_intermediate?: boolean; + file_size: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Name + * @description Name of the model. */ - use_cache?: boolean; + name: string; /** - * @description Latents tensor - * @default null + * Description + * @description Model description */ - latents?: components["schemas"]["LatentsField"] | null; + description: string | null; /** - * @description VAE - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - vae?: components["schemas"]["VAEField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * type - * @default flux2_vae_decode + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter * @constant */ - type: "flux2_vae_decode"; + type: "ip_adapter"; + /** + * Format + * @default invokeai + * @constant + */ + format: "invokeai"; + /** Image Encoder Model Id */ + image_encoder_model_id: string; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; }; - /** - * Image to Latents - FLUX2 - * @description Encodes an image into latents using FLUX.2 Klein's 32-channel VAE. - */ - Flux2VaeEncodeInvocation: { + /** IPAdapter_InvokeAI_SDXL_Config */ + IPAdapter_InvokeAI_SDXL_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * @description The image to encode. - * @default null + * File Size + * @description The size of the model in bytes. */ - image?: components["schemas"]["ImageField"] | null; + file_size: number; /** - * @description VAE - * @default null + * Name + * @description Name of the model. */ - vae?: components["schemas"]["VAEField"] | null; + name: string; /** - * type - * @default flux2_vae_encode + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default ip_adapter * @constant */ - type: "flux2_vae_encode"; + type: "ip_adapter"; + /** + * Format + * @default invokeai + * @constant + */ + format: "invokeai"; + /** Image Encoder Model Id */ + image_encoder_model_id: string; + /** + * Base + * @default sdxl + * @constant + */ + base: "sdxl"; }; /** - * Flux2VariantType - * @description FLUX.2 model variants. - * @enum {string} - */ - Flux2VariantType: "klein_4b" | "klein_9b" | "klein_9b_base"; - /** - * Flux Conditioning Blend - * @description Performs a blend between two FLUX Conditioning objects using either direct SLERP - * or an advanced method that separates magnitude and direction. + * Ideal Size - SD1.5, SDXL + * @description Calculates the ideal size for generation to avoid duplication */ - FluxConditioningBlendInvocation: { + IdealSizeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12887,53 +11778,62 @@ export type components = { */ use_cache?: boolean; /** - * @description The first FLUX Conditioning object. - * @default null + * Width + * @description Final image width + * @default 1024 */ - conditioning_1?: components["schemas"]["FluxConditioningField"] | null; + width?: number; /** - * @description The second FLUX Conditioning object. - * @default null + * Height + * @description Final image height + * @default 576 */ - conditioning_2?: components["schemas"]["FluxConditioningField"] | null; + height?: number; /** - * Alpha - * @description Interpolation factor (0.0 for conditioning_1, 1.0 for conditioning_2). - * @default 0.5 + * @description UNet (scheduler, LoRAs) + * @default null */ - alpha?: number; + unet?: components["schemas"]["UNetField"] | null; /** - * Use Magnitude Separation - * @description If True, uses magnitude separation (SLERP for direction, LERP for magnitude); otherwise, uses direct SLERP. - * @default false + * Multiplier + * @description Amount to multiply the model's dimensions by when calculating the ideal size (may result in initial generation artifacts if too large) + * @default 1 */ - use_magnitude_separation?: boolean; + multiplier?: number; /** * type - * @default flux_conditioning_blend + * @default ideal_size * @constant */ - type: "flux_conditioning_blend"; + type: "ideal_size"; }; /** - * FluxConditioningBlendOutput - * @description Output for the blended Flux Conditionings. + * IdealSizeOutput + * @description Base class for invocations that output an image */ - FluxConditioningBlendOutput: { - /** @description The interpolated Flux Conditioning */ - conditioning: components["schemas"]["FluxConditioningField"]; + IdealSizeOutput: { + /** + * Width + * @description The ideal width of the image (in pixels) + */ + width: number; + /** + * Height + * @description The ideal height of the image (in pixels) + */ + height: number; /** * type - * @default flux_conditioning_blend_output + * @default ideal_size_output * @constant */ - type: "flux_conditioning_blend_output"; + type: "ideal_size_output"; }; /** - * Flux Conditioning Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option + * Image Batch + * @description Create a batched generation, where the workflow is executed once for each image in the batch. */ - FluxConditioningCollectionIndexInvocation: { + ImageBatchInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12948,39 +11848,44 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Random - * @description Random Index? - * @default true + * Batch Group + * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. + * @default None + * @enum {string} */ - random?: boolean; + batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) - * @default 0 - */ - index?: number; - /** - * FLUX Conditionings - * @description Conditioning tensor - * @default [] + * Images + * @description The images to batch over + * @default null */ - collection?: components["schemas"]["FluxConditioningField"][]; + images?: components["schemas"]["ImageField"][] | null; /** * type - * @default flux_conditioning_index + * @default image_batch * @constant */ - type: "flux_conditioning_index"; + type: "image_batch"; }; /** - * Flux Conditioning Collection Primitive - * @description A collection of flux conditioning tensor primitive values + * Blur Image + * @description Blurs an image */ - FluxConditioningCollectionInvocation: { + ImageBlurInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -12999,23 +11904,57 @@ export type components = { */ use_cache?: boolean; /** - * FLUX Conditionings - * @description Conditioning tensor - * @default [] + * @description The image to blur + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Radius + * @description The blur radius + * @default 8 + */ + radius?: number; + /** + * Blur Type + * @description The type of blur + * @default gaussian + * @enum {string} */ - conditioning?: components["schemas"]["FluxConditioningField"][]; + blur_type?: "gaussian" | "box"; /** * type - * @default flux_conditioning_collection + * @default img_blur * @constant */ - type: "flux_conditioning_collection"; + type: "img_blur"; }; /** - * Flux Conditioning Collection join - * @description Join a flux conditioning tensor or collections into a single collection of flux conditioning tensors + * ImageCategory + * @description The category of an image. + * + * - GENERAL: The image is an output, init image, or otherwise an image without a specialized purpose. + * - MASK: The image is a mask image. + * - CONTROL: The image is a ControlNet control image. + * - USER: The image is a user-provide image. + * - OTHER: The image is some other type of image with a specialized purpose. To be used by external nodes. + * @enum {string} + */ + ImageCategory: "general" | "mask" | "control" | "user" | "other"; + /** + * Extract Image Channel + * @description Gets a channel from an image. */ - FluxConditioningCollectionJoinInvocation: { + ImageChannelInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13034,64 +11973,39 @@ export type components = { */ use_cache?: boolean; /** - * FLUX Text Encoder Conditioning or Collection - * @description Conditioning tensor + * @description The image to get the channel from * @default null */ - conditionings_a?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * FLUX Text Encoder Conditioning or Collection - * @description Conditioning tensor - * @default null + * Channel + * @description The channel to get + * @default A + * @enum {string} */ - conditionings_b?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + channel?: "A" | "R" | "G" | "B"; /** * type - * @default flux_conditioning_collection_join + * @default img_chan * @constant */ - type: "flux_conditioning_collection_join"; + type: "img_chan"; }; /** - * FluxConditioningCollectionOutput - * @description Base class for nodes that output a collection of conditioning tensors + * Multiply Image Channel + * @description Scale a specific color channel of an image. */ - FluxConditioningCollectionOutput: { - /** - * Collection - * @description The output conditioning tensors - */ - collection: components["schemas"]["FluxConditioningField"][]; + ImageChannelMultiplyInvocation: { /** - * type - * @default flux_conditioning_collection_output - * @constant + * @description The board to save the image to + * @default null */ - type: "flux_conditioning_collection_output"; - }; - /** - * FluxConditioningDeltaAndAugmentedOutput - * @description Output for the Conditioning Delta and Augmented Conditioning. - */ - FluxConditioningDeltaAndAugmentedOutput: { - /** @description The augmented conditioning (base + delta, or just delta) */ - augmented_conditioning: components["schemas"]["FluxConditioningField"]; - /** @description The resulting conditioning delta (feature - reference) */ - delta_conditioning: components["schemas"]["FluxConditioningField"]; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default conditioning_delta_and_augmented_output - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "conditioning_delta_and_augmented_output"; - }; - /** - * Flux Conditioning Delta - * @description Calculates the delta between feature and reference conditionings, - * and optionally augments a base conditioning with this delta. - * If reference conditioning is omitted, it will be treated as zero tensors. - */ - FluxConditioningDeltaAugmentationInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13110,69 +12024,50 @@ export type components = { */ use_cache?: boolean; /** - * Feature Conditioning - * @description Feature Conditioning (single or list) for delta calculation. If a list, it will be averaged. - * @default null - */ - feature_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; - /** - * Reference Conditioning - * @description Reference Conditioning (single or list) for delta calculation. If a list, it will be averaged. If omitted, zero tensors will be used as reference. + * @description The image to adjust * @default null */ - reference_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description Optional Base Conditioning to which the delta will be added. If not provided, Augmented Conditioning will be the Delta. + * Channel + * @description Which channel to adjust * @default null */ - base_conditioning?: components["schemas"]["FluxConditioningField"] | null; - /** - * Base Scale - * @description Scalar to multiply the base conditioning when augmenting. - * @default 1 - */ - base_scale?: number; + channel?: ("Red (RGBA)" | "Green (RGBA)" | "Blue (RGBA)" | "Alpha (RGBA)" | "Cyan (CMYK)" | "Magenta (CMYK)" | "Yellow (CMYK)" | "Black (CMYK)" | "Hue (HSV)" | "Saturation (HSV)" | "Value (HSV)" | "Luminosity (LAB)" | "A (LAB)" | "B (LAB)" | "Y (YCbCr)" | "Cb (YCbCr)" | "Cr (YCbCr)") | null; /** - * Delta Scale - * @description Scalar to multiply the delta when augmenting the base conditioning. + * Scale + * @description The amount to scale the channel by. * @default 1 */ - delta_scale?: number; + scale?: number; /** - * Scale Delta Output - * @description If true, the delta output will also be scaled by the delta_scale. + * Invert Channel + * @description Invert the channel after scaling * @default false */ - scale_delta_output?: boolean; + invert_channel?: boolean; /** * type - * @default flux_conditioning_delta_augmentation + * @default img_channel_multiply * @constant */ - type: "flux_conditioning_delta_augmentation"; + type: "img_channel_multiply"; }; /** - * FluxConditioningField - * @description A conditioning tensor primitive value + * Offset Image Channel + * @description Add or subtract a value from a specific color channel of an image. */ - FluxConditioningField: { + ImageChannelOffsetInvocation: { /** - * Conditioning Name - * @description The name of conditioning tensor + * @description The board to save the image to + * @default null */ - conditioning_name: string; + board?: components["schemas"]["BoardField"] | null; /** - * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. + * @description Optional metadata to be saved with the image * @default null */ - mask?: components["schemas"]["TensorField"] | null; - }; - /** - * Flux Conditioning List - * @description Takes multiple optional Flux Conditioning inputs and outputs them as a single - * ordered list. Missing (None) inputs are gracefully handled. - */ - FluxConditioningListInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13191,65 +12086,34 @@ export type components = { */ use_cache?: boolean; /** - * @description First optional Flux Conditioning input. - * @default null - */ - conditioning_1?: components["schemas"]["FluxConditioningField"] | null; - /** - * @description Second optional Flux Conditioning input. - * @default null - */ - conditioning_2?: components["schemas"]["FluxConditioningField"] | null; - /** - * @description Third optional Flux Conditioning input. - * @default null - */ - conditioning_3?: components["schemas"]["FluxConditioningField"] | null; - /** - * @description Fourth optional Flux Conditioning input. - * @default null - */ - conditioning_4?: components["schemas"]["FluxConditioningField"] | null; - /** - * @description Fifth optional Flux Conditioning input. + * @description The image to adjust * @default null */ - conditioning_5?: components["schemas"]["FluxConditioningField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description Sixth optional Flux Conditioning input. + * Channel + * @description Which channel to adjust * @default null */ - conditioning_6?: components["schemas"]["FluxConditioningField"] | null; - /** - * type - * @default flux_conditioning_list - * @constant - */ - type: "flux_conditioning_list"; - }; - /** - * FluxConditioningListOutput - * @description Output for the Flux Conditioning List node, providing an ordered list - * of Flux Conditioning objects. - */ - FluxConditioningListOutput: { + channel?: ("Red (RGBA)" | "Green (RGBA)" | "Blue (RGBA)" | "Alpha (RGBA)" | "Cyan (CMYK)" | "Magenta (CMYK)" | "Yellow (CMYK)" | "Black (CMYK)" | "Hue (HSV)" | "Saturation (HSV)" | "Value (HSV)" | "Luminosity (LAB)" | "A (LAB)" | "B (LAB)" | "Y (YCbCr)" | "Cb (YCbCr)" | "Cr (YCbCr)") | null; /** - * Conditioning List - * @description An ordered list of provided Flux Conditioning objects. + * Offset + * @description The amount to adjust the channel by + * @default 0 */ - conditioning_list: components["schemas"]["FluxConditioningField"][]; + offset?: number; /** * type - * @default flux_conditioning_list_output + * @default img_channel_offset * @constant */ - type: "flux_conditioning_list_output"; + type: "img_channel_offset"; }; /** - * FLUX Conditioning Math - * @description Performs a Math operation on two FLUX conditionings. + * Image Collection Primitive + * @description A collection of image primitive values */ - FluxConditioningMathOperationInvocation: { + ImageCollectionInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13268,77 +12132,50 @@ export type components = { */ use_cache?: boolean; /** - * @description First FLUX Conditioning (A) - * @default null - */ - cond_a?: components["schemas"]["FluxConditioningField"] | null; - /** - * @description Second FLUX Conditioning (B) + * Collection + * @description The collection of image values * @default null */ - cond_b?: components["schemas"]["FluxConditioningField"] | null; - /** - * Operation - * @description Operation to perform (A op B) - * @default ADD - * @enum {string} - */ - operation?: "ADD" | "SUB" | "MUL" | "DIV" | "APPEND" | "SPV" | "NSPV" | "LERP" | "SLERP"; - /** - * Scale - * @description Scaling factor - * @default 1 - */ - scale?: number; - /** - * Rescale Target Norm - * @description If > 0, rescales the output embeddings to the target max norm. Set to 0 to disable. - * @default 0 - */ - rescale_target_norm?: number; + collection?: components["schemas"]["ImageField"][] | null; /** * type - * @default flux_conditioning_math + * @default image_collection * @constant */ - type: "flux_conditioning_math"; + type: "image_collection"; }; /** - * FluxConditioningOutput - * @description Base class for nodes that output a single conditioning tensor + * ImageCollectionOutput + * @description Base class for nodes that output a collection of images */ - FluxConditioningOutput: { - /** @description Conditioning tensor */ - conditioning: components["schemas"]["FluxConditioningField"]; + ImageCollectionOutput: { + /** + * Collection + * @description The output images + */ + collection: components["schemas"]["ImageField"][]; /** * type - * @default flux_conditioning_output + * @default image_collection_output * @constant */ - type: "flux_conditioning_output"; + type: "image_collection_output"; }; /** - * FluxConditioningStoreOutput - * @description Output for the Store Flux Conditioning node. + * Convert Image Mode + * @description Converts an image to a different mode. */ - FluxConditioningStoreOutput: { + ImageConvertInvocation: { /** - * Conditioning Id - * @description Unique identifier for the stored Flux Conditioning + * @description The board to save the image to + * @default null */ - conditioning_id: string; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default flux_conditioning_store_output - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "flux_conditioning_store_output"; - }; - /** - * Control LoRA - FLUX - * @description LoRA model and Image to use with FLUX transformer generation. - */ - FluxControlLoRALoaderInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13357,52 +12194,39 @@ export type components = { */ use_cache?: boolean; /** - * Control LoRA - * @description Control LoRA model to load - * @default null - */ - lora?: components["schemas"]["ModelIdentifierField"] | null; - /** - * @description The image to encode. + * @description The image to convert * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Weight - * @description The weight of the LoRA. - * @default 1 + * Mode + * @description The mode to convert to + * @default L + * @enum {string} */ - weight?: number; + mode?: "L" | "RGB" | "RGBA" | "CMYK" | "YCbCr" | "LAB" | "HSV" | "I" | "F"; /** * type - * @default flux_control_lora_loader + * @default img_conv * @constant */ - type: "flux_control_lora_loader"; + type: "img_conv"; }; /** - * FluxControlLoRALoaderOutput - * @description Flux Control LoRA Loader Output + * Crop Image + * @description Crops an image to a specified box. The box can be outside of the image. */ - FluxControlLoRALoaderOutput: { + ImageCropInvocation: { /** - * Flux Control LoRA - * @description Control LoRAs to apply on model loading + * @description The board to save the image to * @default null */ - control_lora: components["schemas"]["ControlLoRAField"]; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default flux_control_lora_loader_output - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "flux_control_lora_loader_output"; - }; - /** - * Flux ControlNet Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option - */ - FluxControlNetCollectionIndexInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13417,167 +12241,141 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Random - * @description Random Index? - * @default true + * @description The image to crop + * @default null */ - random?: boolean; + image?: components["schemas"]["ImageField"] | null; /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) + * X + * @description The left x coordinate of the crop rectangle * @default 0 */ - index?: number; - /** - * FLUX ControlNets - * @description FLUX ControlNet Collection - * @default [] - */ - collection?: components["schemas"]["FluxControlNetField"][]; - /** - * type - * @default flux_controlnet_index - * @constant - */ - type: "flux_controlnet_index"; - }; - /** - * FLUX ControlNet Collection Primitive - * @description A collection of flux controlnet primitive values - */ - FluxControlNetCollectionInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + x?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Y + * @description The top y coordinate of the crop rectangle + * @default 0 */ - is_intermediate?: boolean; + y?: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Width + * @description The width of the crop rectangle + * @default 512 */ - use_cache?: boolean; + width?: number; /** - * FLUX ControlNet Collection - * @description FLUX ControlNets - * @default null + * Height + * @description The height of the crop rectangle + * @default 512 */ - collection?: components["schemas"]["FluxControlNetField"][] | null; + height?: number; /** * type - * @default flux_controlnet_collection + * @default img_crop * @constant */ - type: "flux_controlnet_collection"; + type: "img_crop"; }; /** - * FLUX ControlNet Collection join - * @description Join a flux controlnet tensors or collections into a single collection of flux controlnet tensors + * ImageDTO + * @description Deserialized image record, enriched for the frontend. */ - FluxControlNetCollectionJoinInvocation: { + ImageDTO: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Image Name + * @description The unique name of the image. */ - id: string; + image_name: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Image Url + * @description The URL of the image. */ - is_intermediate?: boolean; + image_url: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Thumbnail Url + * @description The URL of the image's thumbnail. */ - use_cache?: boolean; + thumbnail_url: string; + /** @description The type of the image. */ + image_origin: components["schemas"]["ResourceOrigin"]; + /** @description The category of the image. */ + image_category: components["schemas"]["ImageCategory"]; /** - * FLUX ControlNet or Collection - * @description FLUX ControlNets - * @default null + * Width + * @description The width of the image in px. */ - controlnets_a?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; + width: number; /** - * FLUX ControlNet or Collection - * @description FLUX ControlNets - * @default null + * Height + * @description The height of the image in px. */ - controlnets_b?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; + height: number; /** - * type - * @default flux_controlnet_collection_join - * @constant + * Created At + * @description The created timestamp of the image. */ - type: "flux_controlnet_collection_join"; - }; - /** FluxControlNetCollectionOutput */ - FluxControlNetCollectionOutput: { + created_at: string; /** - * FLUX ControlNet List - * @description ControlNet(s) to apply + * Updated At + * @description The updated timestamp of the image. */ - collection: components["schemas"]["FluxControlNetField"][]; + updated_at: string; /** - * type - * @default flux_controlnet_collection_output - * @constant + * Deleted At + * @description The deleted timestamp of the image. */ - type: "flux_controlnet_collection_output"; - }; - /** FluxControlNetField */ - FluxControlNetField: { - /** @description The control image */ - image: components["schemas"]["ImageField"]; - /** @description The ControlNet model to use */ - control_model: components["schemas"]["ModelIdentifierField"]; + deleted_at?: string | null; /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * Is Intermediate + * @description Whether this is an intermediate image. */ - control_weight?: number | number[]; + is_intermediate: boolean; /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 + * Session Id + * @description The session ID that generated this image, if it is a generated image. */ - begin_step_percent?: number; + session_id?: string | null; /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Node Id + * @description The node ID that generated this image, if it is a generated image. */ - end_step_percent?: number; + node_id?: string | null; /** - * Resize Mode - * @description The resize mode to use - * @default just_resize - * @enum {string} + * Starred + * @description Whether this image is starred. */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + starred: boolean; /** - * Instantx Control Mode - * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. - * @default -1 + * Has Workflow + * @description Whether this image has a workflow. */ - instantx_control_mode?: number | null; + has_workflow: boolean; + /** + * Board Id + * @description The id of the board the image belongs to, if one exists. + */ + board_id?: string | null; }; /** - * FLUX ControlNet - * @description Collect FLUX ControlNet info to pass to other nodes. + * ImageField + * @description An image primitive field */ - FluxControlNetInvocation: { + ImageField: { + /** + * Image Name + * @description The name of the image + */ + image_name: string; + }; + /** + * Image Generator + * @description Generated a collection of images for use in a batched generation + */ + ImageGenerator: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13596,58 +12394,51 @@ export type components = { */ use_cache?: boolean; /** - * @description The control image - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description ControlNet model to load - * @default null - */ - control_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 - */ - control_weight?: number | number[]; - /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 - */ - begin_step_percent?: number; - /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * Generator Type + * @description The image generator. */ - end_step_percent?: number; + generator: components["schemas"]["ImageGeneratorField"]; /** - * Resize Mode - * @description The resize mode used - * @default just_resize - * @enum {string} + * type + * @default image_generator + * @constant */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + type: "image_generator"; + }; + /** ImageGeneratorField */ + ImageGeneratorField: Record; + /** + * ImageGeneratorOutput + * @description Base class for nodes that output a collection of boards + */ + ImageGeneratorOutput: { /** - * Instantx Control Mode - * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. - * @default -1 + * Images + * @description The generated images */ - instantx_control_mode?: number | null; + images: components["schemas"]["ImageField"][]; /** * type - * @default flux_controlnet + * @default image_generator_output * @constant */ - type: "flux_controlnet"; + type: "image_generator_output"; }; /** - * FLUX ControlNet-Linked - * @description Collects FLUX ControlNet info to pass to other nodes. + * Adjust Image Hue + * @description Adjusts the Hue of an image. */ - FluxControlNetLinkedInvocation: { + ImageHueAdjustmentInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13666,92 +12457,84 @@ export type components = { */ use_cache?: boolean; /** - * @description The control image + * @description The image to adjust * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * @description ControlNet model to load - * @default null + * Hue + * @description The degrees by which to rotate the hue, 0-360 + * @default 0 */ - control_model?: components["schemas"]["ModelIdentifierField"] | null; + hue?: number; /** - * Control Weight - * @description The weight given to the ControlNet - * @default 1 + * type + * @default img_hue_adjust + * @constant */ - control_weight?: number | number[]; + type: "img_hue_adjust"; + }; + /** + * Inverse Lerp Image + * @description Inverse linear interpolation of all pixels of an image + */ + ImageInverseLerpInvocation: { /** - * Begin Step Percent - * @description When the ControlNet is first applied (% of total steps) - * @default 0 + * @description The board to save the image to + * @default null */ - begin_step_percent?: number; + board?: components["schemas"]["BoardField"] | null; /** - * End Step Percent - * @description When the ControlNet is last applied (% of total steps) - * @default 1 + * @description Optional metadata to be saved with the image + * @default null */ - end_step_percent?: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Resize Mode - * @description The resize mode used - * @default just_resize - * @enum {string} + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + id: string; /** - * Instantx Control Mode - * @description The control mode for InstantX ControlNet union models. Ignored for other ControlNet models. The standard mapping is: canny (0), tile (1), depth (2), blur (3), pose (4), gray (5), low quality (6). Negative values will be treated as 'None'. - * @default -1 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - instantx_control_mode?: number | null; + is_intermediate?: boolean; /** - * type - * @default flux_controlnet_linked - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - type: "flux_controlnet_linked"; + use_cache?: boolean; /** - * FLUX ControlNet List - * @description FLUX ControlNet List + * @description The image to lerp * @default null */ - controlnet_list?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; - }; - /** FluxControlNetListOutput */ - FluxControlNetListOutput: { + image?: components["schemas"]["ImageField"] | null; /** - * FLUX ControlNet List - * @description ControlNet(s) to apply + * Min + * @description The minimum input value + * @default 0 */ - controlnet_list: components["schemas"]["FluxControlNetField"][]; + min?: number; /** - * type - * @default flux_controlnet_list_output - * @constant + * Max + * @description The maximum input value + * @default 255 */ - type: "flux_controlnet_list_output"; - }; - /** - * FluxControlNetOutput - * @description FLUX ControlNet info - */ - FluxControlNetOutput: { - /** @description ControlNet(s) to apply */ - control: components["schemas"]["FluxControlNetField"]; + max?: number; /** * type - * @default flux_controlnet_output + * @default img_ilerp * @constant */ - type: "flux_controlnet_output"; + type: "img_ilerp"; }; /** - * FLUX Denoise - * @description Run denoising process with a FLUX transformer model. + * Image Primitive + * @description An image primitive value */ - FluxDenoiseInvocation: { + ImageInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -13770,177 +12553,183 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * @description The image to load * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. - * @default null + * type + * @default image + * @constant */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + type: "image"; + }; + /** + * Lerp Image + * @description Linear interpolation of all pixels of an image + */ + ImageLerpInvocation: { /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * @description The board to save the image to + * @default null */ - denoising_start?: number; + board?: components["schemas"]["BoardField"] | null; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * @description Optional metadata to be saved with the image + * @default null */ - denoising_end?: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Add Noise - * @description Add noise based on denoising start. - * @default true + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - add_noise?: boolean; + id: string; /** - * Transformer - * @description Flux model (Transformer) to load - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - transformer?: components["schemas"]["TransformerField"] | null; + is_intermediate?: boolean; /** - * Control LoRA - * @description Control LoRA model to load - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - control_lora?: components["schemas"]["ControlLoRAField"] | null; + use_cache?: boolean; /** - * Positive Text Conditioning - * @description Positive conditioning tensor + * @description The image to lerp * @default null */ - positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Negative Text Conditioning - * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. - * @default null + * Min + * @description The minimum output value + * @default 0 */ - negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + min?: number; /** - * Redux Conditioning - * @description FLUX Redux conditioning tensor. - * @default null + * Max + * @description The maximum output value + * @default 255 */ - redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; + max?: number; /** - * @description FLUX Fill conditioning. - * @default null + * type + * @default img_lerp + * @constant */ - fill_conditioning?: components["schemas"]["FluxFillConditioningField"] | null; + type: "img_lerp"; + }; + /** + * Image Mask to Tensor + * @description Convert a mask image to a tensor. Converts the image to grayscale and uses thresholding at the specified value. + */ + ImageMaskToTensorInvocation: { /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 1 + * @description Optional metadata to be saved with the image + * @default null */ - cfg_scale?: number | number[]; + metadata?: components["schemas"]["MetadataField"] | null; /** - * CFG Scale Start Step - * @description Index of the first step to apply cfg_scale. Negative indices count backwards from the the last step (e.g. a value of -1 refers to the final step). - * @default 0 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - cfg_scale_start_step?: number; + id: string; /** - * CFG Scale End Step - * @description Index of the last step to apply cfg_scale. Negative indices count backwards from the last step (e.g. a value of -1 refers to the final step). - * @default -1 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cfg_scale_end_step?: number; + is_intermediate?: boolean; /** - * Width - * @description Width of the generated image. - * @default 1024 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - width?: number; + use_cache?: boolean; /** - * Height - * @description Height of the generated image. - * @default 1024 + * @description The mask image to convert. + * @default null */ - height?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Num Steps - * @description Number of diffusion steps. Recommended values are schnell: 4, dev: 50. - * @default 4 + * Cutoff + * @description Cutoff (<) + * @default 128 */ - num_steps?: number; + cutoff?: number; /** - * Scheduler - * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. - * @default euler - * @enum {string} + * Invert + * @description Whether to invert the mask. + * @default false */ - scheduler?: "euler" | "heun" | "lcm"; + invert?: boolean; /** - * Guidance - * @description The guidance strength. Higher values adhere more strictly to the prompt, and will produce less diverse images. FLUX dev only, ignored for schnell. - * @default 4 + * type + * @default image_mask_to_tensor + * @constant */ - guidance?: number; + type: "image_mask_to_tensor"; + }; + /** + * Multiply Images + * @description Multiplies two images together using `PIL.ImageChops.multiply()`. + */ + ImageMultiplyInvocation: { /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * @description The board to save the image to + * @default null */ - seed?: number; + board?: components["schemas"]["BoardField"] | null; /** - * Control - * @description ControlNet models. + * @description Optional metadata to be saved with the image * @default null */ - control?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; - /** - * @description VAE - * @default null - */ - controlnet_vae?: components["schemas"]["VAEField"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * IP-Adapter - * @description IP-Adapter to apply - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + id: string; /** - * Kontext Conditioning - * @description FLUX Kontext conditioning (reference image). - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; + is_intermediate?: boolean; /** - * Dype Preset - * @description DyPE preset for high-resolution generation. 'auto' enables automatically for resolutions > 1536px. 'area' enables automatically based on image area. '4k' uses optimized settings for 4K output. - * @default off - * @enum {string} + * Use Cache + * @description Whether or not to use the cache + * @default true */ - dype_preset?: "off" | "manual" | "auto" | "area" | "4k"; + use_cache?: boolean; /** - * Dype Scale - * @description DyPE magnitude (λs). Higher values = stronger extrapolation. Only used when dype_preset is not 'off'. + * @description The first image to multiply * @default null */ - dype_scale?: number | null; + image1?: components["schemas"]["ImageField"] | null; /** - * Dype Exponent - * @description DyPE decay speed (λt). Controls transition from low to high frequency detail. Only used when dype_preset is not 'off'. + * @description The second image to multiply * @default null */ - dype_exponent?: number | null; + image2?: components["schemas"]["ImageField"] | null; /** * type - * @default flux_denoise + * @default img_mul * @constant */ - type: "flux_denoise"; + type: "img_mul"; }; /** - * FLUX Denoise + Metadata - * @description Run denoising process with a FLUX transformer model + metadata. + * Blur NSFW Image + * @description Add blur to NSFW-flagged images */ - FluxDenoiseLatentsMetaInvocation: { + ImageNSFWBlurInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -13964,187 +12753,177 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * @description The image to check * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. - * @default null + * type + * @default img_nsfw + * @constant */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + type: "img_nsfw"; + }; + /** + * ImageNamesResult + * @description Response containing ordered image names with metadata for optimistic updates. + */ + ImageNamesResult: { /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Image Names + * @description Ordered list of image names */ - denoising_start?: number; + image_names: string[]; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * Starred Count + * @description Number of starred images (when starred_first=True) */ - denoising_end?: number; + starred_count: number; /** - * Add Noise - * @description Add noise based on denoising start. - * @default true + * Total Count + * @description Total number of images matching the query */ - add_noise?: boolean; + total_count: number; + }; + /** + * Add Image Noise + * @description Add noise to an image + */ + ImageNoiseInvocation: { /** - * Transformer - * @description Flux model (Transformer) to load + * @description The board to save the image to * @default null */ - transformer?: components["schemas"]["TransformerField"] | null; + board?: components["schemas"]["BoardField"] | null; /** - * Control LoRA - * @description Control LoRA model to load + * @description Optional metadata to be saved with the image * @default null */ - control_lora?: components["schemas"]["ControlLoRAField"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Positive Text Conditioning - * @description Positive conditioning tensor - * @default null + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - positive_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + id: string; /** - * Negative Text Conditioning - * @description Negative conditioning tensor. Can be None if cfg_scale is 1.0. - * @default null + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - negative_text_conditioning?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; + is_intermediate?: boolean; /** - * Redux Conditioning - * @description FLUX Redux conditioning tensor. - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default true */ - redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; + use_cache?: boolean; /** - * @description FLUX Fill conditioning. + * @description The image to add noise to * @default null */ - fill_conditioning?: components["schemas"]["FluxFillConditioningField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 1 + * @description Optional mask determining where to apply noise (black=noise, white=no noise) + * @default null */ - cfg_scale?: number | number[]; + mask?: components["schemas"]["ImageField"] | null; /** - * CFG Scale Start Step - * @description Index of the first step to apply cfg_scale. Negative indices count backwards from the the last step (e.g. a value of -1 refers to the final step). + * Seed + * @description Seed for random number generation * @default 0 */ - cfg_scale_start_step?: number; + seed?: number; /** - * CFG Scale End Step - * @description Index of the last step to apply cfg_scale. Negative indices count backwards from the last step (e.g. a value of -1 refers to the final step). - * @default -1 + * Noise Type + * @description The type of noise to add + * @default gaussian + * @enum {string} */ - cfg_scale_end_step?: number; + noise_type?: "gaussian" | "salt_and_pepper"; /** - * Width - * @description Width of the generated image. - * @default 1024 + * Amount + * @description The amount of noise to add + * @default 0.1 */ - width?: number; + amount?: number; /** - * Height - * @description Height of the generated image. - * @default 1024 + * Noise Color + * @description Whether to add colored noise + * @default true */ - height?: number; + noise_color?: boolean; /** - * Num Steps - * @description Number of diffusion steps. Recommended values are schnell: 4, dev: 50. - * @default 4 + * Size + * @description The size of the noise points + * @default 1 */ - num_steps?: number; + size?: number; /** - * Scheduler - * @description Scheduler (sampler) for the denoising process. 'euler' is fast and standard. 'heun' is 2nd-order (better quality, 2x slower). 'lcm' is optimized for few steps. - * @default euler - * @enum {string} + * type + * @default img_noise + * @constant */ - scheduler?: "euler" | "heun" | "lcm"; + type: "img_noise"; + }; + /** + * ImageOutput + * @description Base class for nodes that output a single image + */ + ImageOutput: { + /** @description The output image */ + image: components["schemas"]["ImageField"]; /** - * Guidance - * @description The guidance strength. Higher values adhere more strictly to the prompt, and will produce less diverse images. FLUX dev only, ignored for schnell. - * @default 4 + * Width + * @description The width of the image in pixels */ - guidance?: number; + width: number; /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * Height + * @description The height of the image in pixels */ - seed?: number; + height: number; /** - * Control - * @description ControlNet models. - * @default null + * type + * @default image_output + * @constant */ - control?: components["schemas"]["FluxControlNetField"] | components["schemas"]["FluxControlNetField"][] | null; + type: "image_output"; + }; + /** ImagePanelCoordinateOutput */ + ImagePanelCoordinateOutput: { /** - * @description VAE - * @default null + * X Left + * @description The left x-coordinate of the panel. */ - controlnet_vae?: components["schemas"]["VAEField"] | null; + x_left: number; /** - * IP-Adapter - * @description IP-Adapter to apply - * @default null + * Y Top + * @description The top y-coordinate of the panel. */ - ip_adapter?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + y_top: number; /** - * Kontext Conditioning - * @description FLUX Kontext conditioning (reference image). - * @default null + * Width + * @description The width of the panel. */ - kontext_conditioning?: components["schemas"]["FluxKontextConditioningField"] | components["schemas"]["FluxKontextConditioningField"][] | null; + width: number; /** - * Dype Preset - * @description DyPE preset for high-resolution generation. 'auto' enables automatically for resolutions > 1536px. 'area' enables automatically based on image area. '4k' uses optimized settings for 4K output. - * @default off - * @enum {string} + * Height + * @description The height of the panel. */ - dype_preset?: "off" | "manual" | "auto" | "area" | "4k"; - /** - * Dype Scale - * @description DyPE magnitude (λs). Higher values = stronger extrapolation. Only used when dype_preset is not 'off'. - * @default null - */ - dype_scale?: number | null; - /** - * Dype Exponent - * @description DyPE decay speed (λt). Controls transition from low to high frequency detail. Only used when dype_preset is not 'off'. - * @default null - */ - dype_exponent?: number | null; + height: number; /** * type - * @default flux_denoise_meta + * @default image_panel_coordinate_output * @constant */ - type: "flux_denoise_meta"; - }; - /** - * FluxFillConditioningField - * @description A FLUX Fill conditioning field. - */ - FluxFillConditioningField: { - /** @description The FLUX Fill reference image. */ - image: components["schemas"]["ImageField"]; - /** @description The FLUX Fill inpaint mask. */ - mask: components["schemas"]["TensorField"]; + type: "image_panel_coordinate_output"; }; /** - * FLUX Fill Conditioning - * @description Prepare the FLUX Fill conditioning data. + * Image Panel Layout + * @description Get the coordinates of a single panel in a grid. (If the full image shape cannot be divided evenly into panels, + * then the grid may not cover the entire image.) */ - FluxFillInvocation: { + ImagePanelLayoutInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14163,44 +12942,63 @@ export type components = { */ use_cache?: boolean; /** - * @description The FLUX Fill reference image. + * Width + * @description The width of the entire grid. * @default null */ - image?: components["schemas"]["ImageField"] | null; + width?: number | null; /** - * @description The bool inpainting mask. Excluded regions should be set to False, included regions should be set to True. + * Height + * @description The height of the entire grid. * @default null */ - mask?: components["schemas"]["TensorField"] | null; + height?: number | null; /** - * type - * @default flux_fill - * @constant + * Num Cols + * @description The number of columns in the grid. + * @default 1 */ - type: "flux_fill"; - }; - /** - * FluxFillOutput - * @description The conditioning output of a FLUX Fill invocation. - */ - FluxFillOutput: { + num_cols?: number; /** - * Conditioning - * @description FLUX Redux conditioning tensor + * Num Rows + * @description The number of rows in the grid. + * @default 1 */ - fill_cond: components["schemas"]["FluxFillConditioningField"]; + num_rows?: number; + /** + * Panel Col Idx + * @description The column index of the panel to be processed. + * @default 0 + */ + panel_col_idx?: number; + /** + * Panel Row Idx + * @description The row index of the panel to be processed. + * @default 0 + */ + panel_row_idx?: number; /** * type - * @default flux_fill_output + * @default image_panel_layout * @constant */ - type: "flux_fill_output"; + type: "image_panel_layout"; }; /** - * FLUX IP-Adapter - * @description Collects FLUX IP-Adapter info to pass to other nodes. + * Paste Image + * @description Pastes an image into another image. */ - FluxIPAdapterInvocation: { + ImagePasteInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14219,53 +13017,91 @@ export type components = { */ use_cache?: boolean; /** - * @description The IP-Adapter image prompt(s). + * @description The base image * @default null */ - image?: components["schemas"]["ImageField"] | null; + base_image?: components["schemas"]["ImageField"] | null; /** - * IP-Adapter Model - * @description The IP-Adapter model. + * @description The image to paste * @default null */ - ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Clip Vision Model - * @description CLIP Vision model to use. - * @default ViT-L - * @constant + * @description The mask to use when pasting + * @default null */ - clip_vision_model?: "ViT-L"; + mask?: components["schemas"]["ImageField"] | null; /** - * Weight - * @description The weight given to the IP-Adapter - * @default 1 + * X + * @description The left x coordinate at which to paste the image + * @default 0 */ - weight?: number | number[]; + x?: number; /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) + * Y + * @description The top y coordinate at which to paste the image * @default 0 */ - begin_step_percent?: number; + y?: number; /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) - * @default 1 + * Crop + * @description Crop to base image dimensions + * @default false */ - end_step_percent?: number; + crop?: boolean; /** * type - * @default flux_ip_adapter + * @default img_paste * @constant */ - type: "flux_ip_adapter"; + type: "img_paste"; }; /** - * Flux Ideal Size - * @description Calculates the ideal size for generation to avoid duplication + * ImageRecordChanges + * @description A set of changes to apply to an image record. + * + * Only limited changes are valid: + * - `image_category`: change the category of an image + * - `session_id`: change the session associated with an image + * - `is_intermediate`: change the image's `is_intermediate` flag + * - `starred`: change whether the image is starred + */ + ImageRecordChanges: { + /** @description The image's new category. */ + image_category?: components["schemas"]["ImageCategory"] | null; + /** + * Session Id + * @description The image's new session ID. + */ + session_id?: string | null; + /** + * Is Intermediate + * @description The image's new `is_intermediate` flag. + */ + is_intermediate?: boolean | null; + /** + * Starred + * @description The image's new `starred` state + */ + starred?: boolean | null; + } & { + [key: string]: unknown; + }; + /** + * Resize Image + * @description Resizes an image to specific dimensions */ - FluxIdealSizeInvocation: { + ImageResizeInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14283,59 +13119,42 @@ export type components = { * @default true */ use_cache?: boolean; + /** + * @description The image to resize + * @default null + */ + image?: components["schemas"]["ImageField"] | null; /** * Width - * @description Target width - * @default 1024 + * @description The width to resize to (px) + * @default 512 */ width?: number; /** * Height - * @description Target height - * @default 576 + * @description The height to resize to (px) + * @default 512 */ height?: number; /** - * Multiplier - * @description Dimensional multiplier - * @default 1 - */ - multiplier?: number; - /** - * type - * @default flux_ideal_size - * @constant - */ - type: "flux_ideal_size"; - }; - /** - * FluxIdealSizeOutput - * @description Base class for invocations that output an image - */ - FluxIdealSizeOutput: { - /** - * Width - * @description The ideal width of the image in pixels - */ - width: number; - /** - * Height - * @description The ideal height of the image in pixels + * Resample Mode + * @description The resampling mode + * @default bicubic + * @enum {string} */ - height: number; + resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; /** * type - * @default flux_ideal_size_output + * @default img_resize * @constant */ - type: "flux_ideal_size_output"; + type: "img_resize"; }; /** - * FLUX Kontext Image Prep - * @description Prepares an image or images for use with FLUX Kontext. The first/single image is resized to the nearest - * preferred Kontext resolution. All other images are concatenated horizontally, maintaining their aspect ratio. + * Scale Image + * @description Scales an image by a factor */ - FluxKontextConcatenateImagesInvocation: { + ImageScaleInvocation: { /** * @description The board to save the image to * @default null @@ -14364,37 +13183,35 @@ export type components = { */ use_cache?: boolean; /** - * Images - * @description The images to concatenate + * @description The image to scale * @default null */ - images?: components["schemas"]["ImageField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Use Preferred Resolution - * @description Use FLUX preferred resolutions for the first image - * @default true + * Scale Factor + * @description The factor by which to scale the image + * @default 2 */ - use_preferred_resolution?: boolean; + scale_factor?: number; + /** + * Resample Mode + * @description The resampling mode + * @default bicubic + * @enum {string} + */ + resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; /** * type - * @default flux_kontext_image_prep + * @default img_scale * @constant */ - type: "flux_kontext_image_prep"; + type: "img_scale"; }; /** - * FluxKontextConditioningField - * @description A conditioning field for FLUX Kontext (reference image). + * Image to Latents - SD1.5, SDXL + * @description Encodes an image into latents. */ - FluxKontextConditioningField: { - /** @description The Kontext reference image. */ - image: components["schemas"]["ImageField"]; - }; - /** - * Flux Kontext Ideal Size - * @description Calculates the ideal size for generation using Flux Kontext - */ - FluxKontextIdealSizeInvocation: { + ImageToLatentsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14413,29 +13230,93 @@ export type components = { */ use_cache?: boolean; /** - * Width - * @description Target width - * @default 1024 + * @description The image to encode + * @default null */ - width?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Height - * @description Target height - * @default 576 + * @description VAE + * @default null */ - height?: number; + vae?: components["schemas"]["VAEField"] | null; + /** + * Tiled + * @description Processing using overlapping tiles (reduce memory consumption) + * @default false + */ + tiled?: boolean; + /** + * Tile Size + * @description The tile size for VAE tiling in pixels (image space). If set to 0, the default tile size for the model will be used. Larger tile sizes generally produce better results at the cost of higher memory usage. + * @default 0 + */ + tile_size?: number; + /** + * Fp32 + * @description Whether or not to use full float32 precision + * @default false + */ + fp32?: boolean; + /** + * Color Compensation + * @description Apply VAE scaling compensation when encoding images (reduces color drift). + * @default None + * @enum {string} + */ + color_compensation?: "None" | "SDXL"; /** * type - * @default flux_kontext_ideal_size + * @default i2l * @constant */ - type: "flux_kontext_ideal_size"; + type: "i2l"; + }; + /** ImageUploadEntry */ + ImageUploadEntry: { + /** @description The image DTO */ + image_dto: components["schemas"]["ImageDTO"]; + /** + * Presigned Url + * @description The URL to get the presigned URL for the image upload + */ + presigned_url: string; }; /** - * Kontext Conditioning - FLUX - * @description Prepares a reference image for FLUX Kontext conditioning. + * ImageUrlsDTO + * @description The URLs for an image and its thumbnail. */ - FluxKontextInvocation: { + ImageUrlsDTO: { + /** + * Image Name + * @description The unique name of the image. + */ + image_name: string; + /** + * Image Url + * @description The URL of the image. + */ + image_url: string; + /** + * Thumbnail Url + * @description The URL of the image's thumbnail. + */ + thumbnail_url: string; + }; + /** + * Add Invisible Watermark + * @description Add an invisible watermark to an image + */ + ImageWatermarkInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14454,39 +13335,51 @@ export type components = { */ use_cache?: boolean; /** - * @description The Kontext reference image. + * @description The image to check * @default null */ image?: components["schemas"]["ImageField"] | null; + /** + * Text + * @description Watermark text + * @default InvokeAI + */ + text?: string; /** * type - * @default flux_kontext + * @default img_watermark * @constant */ - type: "flux_kontext"; + type: "img_watermark"; }; - /** - * FluxKontextOutput - * @description The conditioning output of a FLUX Kontext invocation. - */ - FluxKontextOutput: { + /** ImagesDownloaded */ + ImagesDownloaded: { /** - * Kontext Conditioning - * @description FLUX Kontext conditioning (reference image) + * Response + * @description The message to display to the user when images begin downloading */ - kontext_cond: components["schemas"]["FluxKontextConditioningField"]; + response?: string | null; /** - * type - * @default flux_kontext_output - * @constant + * Bulk Download Item Name + * @description The name of the bulk download item for which events will be emitted */ - type: "flux_kontext_output"; + bulk_download_item_name?: string | null; }; /** - * Apply LoRA - FLUX - * @description Apply a LoRA model to a FLUX transformer and/or text encoder. + * Solid Color Infill + * @description Infills transparent areas of an image with a solid color */ - FluxLoRALoaderInvocation: { + InfillColorInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14505,77 +13398,42 @@ export type components = { */ use_cache?: boolean; /** - * LoRA - * @description LoRA model to load - * @default null - */ - lora?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 - */ - weight?: number; - /** - * FLUX Transformer - * @description Transformer - * @default null - */ - transformer?: components["schemas"]["TransformerField"] | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @description The image to process * @default null */ - clip?: components["schemas"]["CLIPField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * T5 Encoder - * @description T5 tokenizer and text encoder - * @default null + * @description The color to use to infill + * @default { + * "r": 127, + * "g": 127, + * "b": 127, + * "a": 255 + * } */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; + color?: components["schemas"]["ColorField"]; /** * type - * @default flux_lora_loader + * @default infill_rgba * @constant */ - type: "flux_lora_loader"; + type: "infill_rgba"; }; /** - * FluxLoRALoaderOutput - * @description FLUX LoRA Loader Output + * PatchMatch Infill + * @description Infills transparent areas of an image using the PatchMatch algorithm */ - FluxLoRALoaderOutput: { - /** - * FLUX Transformer - * @description Transformer - * @default null - */ - transformer: components["schemas"]["TransformerField"] | null; + InfillPatchMatchInvocation: { /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @description The board to save the image to * @default null */ - clip: components["schemas"]["CLIPField"] | null; + board?: components["schemas"]["BoardField"] | null; /** - * T5 Encoder - * @description T5 tokenizer and text encoder + * @description Optional metadata to be saved with the image * @default null */ - t5_encoder: components["schemas"]["T5EncoderField"] | null; - /** - * type - * @default flux_lora_loader_output - * @constant - */ - type: "flux_lora_loader_output"; - }; - /** - * FLUX Main Model Input - * @description Loads a flux model from an input, outputting its submodels. - */ - FluxModelLoaderInputInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14594,40 +13452,45 @@ export type components = { */ use_cache?: boolean; /** - * @description Flux model (Transformer) to load - * @default null - */ - model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * T5 Encoder - * @description T5 tokenizer and text encoder + * @description The image to process * @default null */ - t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * CLIP Embed - * @description CLIP Embed loader - * @default null + * Downscale + * @description Run patchmatch on downscaled image to speedup infill + * @default 2 */ - clip_embed_model?: components["schemas"]["ModelIdentifierField"] | null; + downscale?: number; /** - * VAE - * @description VAE model to load - * @default null + * Resample Mode + * @description The resampling mode + * @default bicubic + * @enum {string} */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; /** * type - * @default flux_model_loader_input + * @default infill_patchmatch * @constant */ - type: "flux_model_loader_input"; + type: "infill_patchmatch"; }; /** - * Main Model - FLUX - * @description Loads a flux base model, outputting its submodels. + * Tile Infill + * @description Infills transparent areas of an image with tiles of the image */ - FluxModelLoaderInvocation: { + InfillTileInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14646,78 +13509,114 @@ export type components = { */ use_cache?: boolean; /** - * @description Flux model (Transformer) to load + * @description The image to process * @default null */ - model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * T5 Encoder - * @description T5 tokenizer and text encoder - * @default null - */ - t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * CLIP Embed - * @description CLIP Embed loader - * @default null + * Tile Size + * @description The tile size (px) + * @default 32 */ - clip_embed_model?: components["schemas"]["ModelIdentifierField"] | null; + tile_size?: number; /** - * VAE - * @description VAE model to load - * @default null + * Seed + * @description The seed to use for tile generation (omit for random) + * @default 0 */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + seed?: number; /** * type - * @default flux_model_loader + * @default infill_tile * @constant */ - type: "flux_model_loader"; + type: "infill_tile"; }; /** - * FluxModelLoaderOutput - * @description Flux base model loader output + * Input + * @description The type of input a field accepts. + * - `Input.Direct`: The field must have its value provided directly, when the invocation and field are instantiated. + * - `Input.Connection`: The field must have its value provided by a connection. + * - `Input.Any`: The field may have its value provided either directly or by a connection. + * @enum {string} */ - FluxModelLoaderOutput: { + Input: "connection" | "direct" | "any"; + /** + * InputFieldJSONSchemaExtra + * @description Extra attributes to be added to input fields and their OpenAPI schema. Used during graph execution, + * and by the workflow editor during schema parsing and UI rendering. + */ + InputFieldJSONSchemaExtra: { + input: components["schemas"]["Input"]; + field_kind: components["schemas"]["FieldKind"]; /** - * Transformer - * @description Transformer + * Orig Required + * @default true */ - transformer: components["schemas"]["TransformerField"]; + orig_required: boolean; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Default + * @default null */ - clip: components["schemas"]["CLIPField"]; + default: unknown | null; /** - * T5 Encoder - * @description T5 tokenizer and text encoder + * Orig Default + * @default null */ - t5_encoder: components["schemas"]["T5EncoderField"]; + orig_default: unknown | null; /** - * VAE - * @description VAE + * Ui Hidden + * @default false */ - vae: components["schemas"]["VAEField"]; + ui_hidden: boolean; + /** @default null */ + ui_type: components["schemas"]["UIType"] | null; + /** @default null */ + ui_component: components["schemas"]["UIComponent"] | null; /** - * Max Seq Length - * @description The max sequence length to used for the T5 encoder. (256 for schnell transformer, 512 for dev transformer) - * @enum {integer} + * Ui Order + * @default null */ - max_seq_len: 256 | 512; + ui_order: number | null; /** - * type - * @default flux_model_loader_output - * @constant + * Ui Choice Labels + * @default null */ - type: "flux_model_loader_output"; + ui_choice_labels: { + [key: string]: string; + } | null; + /** + * Ui Model Base + * @default null + */ + ui_model_base: components["schemas"]["BaseModelType"][] | null; + /** + * Ui Model Type + * @default null + */ + ui_model_type: components["schemas"]["ModelType"][] | null; + /** + * Ui Model Variant + * @default null + */ + ui_model_variant: (components["schemas"]["ClipVariantType"] | components["schemas"]["ModelVariantType"])[] | null; + /** + * Ui Model Format + * @default null + */ + ui_model_format: components["schemas"]["ModelFormat"][] | null; }; /** - * Flux Model To String - * @description Converts an Flux Model to a JSONString + * InstallStatus + * @description State of an install job running in the background. + * @enum {string} + */ + InstallStatus: "waiting" | "downloading" | "downloads_done" | "running" | "paused" | "completed" | "error" | "cancelled"; + /** + * Integer Batch + * @description Create a batched generation, where the workflow is executed once for each integer in the batch. */ - FluxModelToStringInvocation: { + IntegerBatchInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14736,22 +13635,30 @@ export type components = { */ use_cache?: boolean; /** - * @description Flux model (Transformer) to load + * Batch Group + * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. + * @default None + * @enum {string} + */ + batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; + /** + * Integers + * @description The integers to batch over * @default null */ - model?: components["schemas"]["ModelIdentifierField"] | null; + integers?: number[] | null; /** * type - * @default flux_model_to_string + * @default integer_batch * @constant */ - type: "flux_model_to_string"; + type: "integer_batch"; }; /** - * Flux Redux Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option + * Integer Collection Primitive + * @description A collection of integer primitive values */ - FluxReduxCollectionIndexInvocation: { + IntegerCollectionInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14766,39 +13673,44 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Random - * @description Random Index? - * @default true + * Collection + * @description The collection of integer values + * @default [] */ - random?: boolean; + collection?: number[]; /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) - * @default 0 + * type + * @default integer_collection + * @constant */ - index?: number; + type: "integer_collection"; + }; + /** + * IntegerCollectionOutput + * @description Base class for nodes that output a collection of integers + */ + IntegerCollectionOutput: { /** - * FLUX Redux Collection - * @description Conditioning tensor - * @default [] + * Collection + * @description The int collection */ - collection?: components["schemas"]["FluxReduxConditioningField"][]; + collection: number[]; /** * type - * @default flux_redux_index + * @default integer_collection_output * @constant */ - type: "flux_redux_index"; + type: "integer_collection_output"; }; /** - * FLUX Redux Collection Primitive - * @description A collection of flux redux primitive values + * Integer Generator + * @description Generated a range of integers for use in a batched generation */ - FluxReduxCollectionInvocation: { + IntegerGenerator: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14817,23 +13729,38 @@ export type components = { */ use_cache?: boolean; /** - * FLUX Redux Collection - * @description FLUX Redux Collection - * @default null + * Generator Type + * @description The integer generator. + */ + generator: components["schemas"]["IntegerGeneratorField"]; + /** + * type + * @default integer_generator + * @constant + */ + type: "integer_generator"; + }; + /** IntegerGeneratorField */ + IntegerGeneratorField: Record; + /** IntegerGeneratorOutput */ + IntegerGeneratorOutput: { + /** + * Integers + * @description The generated integers */ - collection?: components["schemas"]["FluxReduxConditioningField"][] | null; + integers: number[]; /** * type - * @default flux_redux_collection + * @default integer_generator_output * @constant */ - type: "flux_redux_collection"; + type: "integer_generator_output"; }; /** - * FLUX Redux Collection join - * @description Join a flux redux tensor or collections into a single collection of flux redux tensors + * Integer Primitive + * @description An integer primitive value */ - FluxReduxCollectionJoinInvocation: { + IntegerInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14852,56 +13779,23 @@ export type components = { */ use_cache?: boolean; /** - * FLUX Redux or Collection - * @description FLUX Reduxs - * @default null - */ - conditionings_a?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; - /** - * FLUX Redux or Collection - * @description FLUX Reduxs - * @default null - */ - conditionings_b?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; - /** - * type - * @default flux_redux_collection_join - * @constant - */ - type: "flux_redux_collection_join"; - }; - /** FluxReduxCollectionOutput */ - FluxReduxCollectionOutput: { - /** - * FLUX Redux Collection - * @description ControlNet(s) to apply + * Value + * @description The integer value + * @default 0 */ - collection: components["schemas"]["FluxReduxConditioningField"][]; + value?: number; /** * type - * @default flux_redux_collection_output + * @default integer * @constant */ - type: "flux_redux_collection_output"; - }; - /** - * FluxReduxConditioningField - * @description A FLUX Redux conditioning tensor primitive value - */ - FluxReduxConditioningField: { - /** @description The Redux image conditioning tensor. */ - conditioning: components["schemas"]["TensorField"]; - /** - * @description The mask associated with this conditioning tensor. Excluded regions should be set to False, included regions should be set to True. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; + type: "integer"; }; /** - * FLUX Redux Conditioning Math - * @description Performs a Math operation on two FLUX Redux conditionings. + * Integer Math + * @description Performs integer math. */ - FluxReduxConditioningMathOperationInvocation: { + IntegerMathInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -14919,100 +13813,54 @@ export type components = { * @default true */ use_cache?: boolean; - /** - * @description First FLUX Redux Conditioning (A) - * @default null - */ - cond_a?: components["schemas"]["FluxReduxConditioningField"] | null; - /** - * @description Second FLUX Redux Conditioning (B) - * @default null - */ - cond_b?: components["schemas"]["FluxReduxConditioningField"] | null; /** * Operation - * @description Operation to perform (A op B) + * @description The operation to perform * @default ADD * @enum {string} */ - operation?: "ADD" | "SUB" | "MUL" | "DIV" | "APPEND" | "SPV" | "NSPV" | "LERP" | "SLERP"; + operation?: "ADD" | "SUB" | "MUL" | "DIV" | "EXP" | "MOD" | "ABS" | "MIN" | "MAX"; /** - * Scale - * @description Scaling factor + * A + * @description The first number * @default 1 */ - scale?: number; + a?: number; /** - * Rescale Target Norm - * @description If > 0, rescales the output embeddings to the target max norm. Set to 0 to disable. - * @default 0 + * B + * @description The second number + * @default 1 */ - rescale_target_norm?: number; + b?: number; /** * type - * @default flux_redux_conditioning_math + * @default integer_math * @constant */ - type: "flux_redux_conditioning_math"; + type: "integer_math"; }; /** - * FLUX Redux Downsampling - * @description Downsampling Flux Redux conditioning + * IntegerOutput + * @description Base class for nodes that output a single integer */ - FluxReduxDownsamplingInvocation: { + IntegerOutput: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Value + * @description The output integer */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description FLUX Redux conditioning tensor. - * @default null - */ - redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | null; - /** - * Downsampling Factor - * @description Redux Downsampling Factor (1-9) - * @default 3 - */ - downsampling_factor?: number; - /** - * Downsampling Function - * @description Redux Downsampling Function - * @default area - * @enum {string} - */ - downsampling_function?: "nearest" | "bilinear" | "bicubic" | "area" | "nearest-exact"; - /** - * Weight - * @description Redux weight (0.0-3.0) - * @default 1 - */ - weight?: number; + value: number; /** * type - * @default flux_redux_downsampling + * @default integer_output * @constant */ - type: "flux_redux_downsampling"; + type: "integer_output"; }; /** - * FLUX Redux - * @description Runs a FLUX Redux model to generate a conditioning tensor. + * Invert Tensor Mask + * @description Inverts a tensor mask. */ - FluxReduxInvocation: { + InvertTensorMaskInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -15031,14846 +13879,1019 @@ export type components = { */ use_cache?: boolean; /** - * @description The FLUX Redux image prompt. + * @description The tensor mask to convert. * @default null */ - image?: components["schemas"]["ImageField"] | null; + mask?: components["schemas"]["TensorField"] | null; /** - * @description The bool mask associated with this FLUX Redux image prompt. Excluded regions should be set to False, included regions should be set to True. - * @default null + * type + * @default invert_tensor_mask + * @constant */ - mask?: components["schemas"]["TensorField"] | null; + type: "invert_tensor_mask"; + }; + /** InvocationCacheStatus */ + InvocationCacheStatus: { /** - * FLUX Redux Model - * @description The FLUX Redux model to use. - * @default null + * Size + * @description The current size of the invocation cache */ - redux_model?: components["schemas"]["ModelIdentifierField"] | null; + size: number; /** - * Downsampling Factor - * @description Redux Downsampling Factor (1-9) - * @default 1 + * Hits + * @description The number of cache hits */ - downsampling_factor?: number; + hits: number; /** - * Downsampling Function - * @description Redux Downsampling Function - * @default area - * @enum {string} + * Misses + * @description The number of cache misses */ - downsampling_function?: "nearest" | "bilinear" | "bicubic" | "area" | "nearest-exact"; + misses: number; /** - * Weight - * @description Redux weight (0.0-1.0) - * @default 1 + * Enabled + * @description Whether the invocation cache is enabled */ - weight?: number; + enabled: boolean; /** - * type - * @default flux_redux - * @constant + * Max Size + * @description The maximum size of the invocation cache */ - type: "flux_redux"; + max_size: number; }; /** - * FLUX Redux-Linked - * @description Collects FLUX Redux conditioning info to pass to other nodes. + * InvocationCompleteEvent + * @description Event model for invocation_complete */ - FluxReduxLinkedInvocation: { + InvocationCompleteEvent: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Timestamp + * @description The timestamp of the event */ - id: string; + timestamp: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Queue Id + * @description The ID of the queue */ - is_intermediate?: boolean; + queue_id: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Item Id + * @description The ID of the queue item */ - use_cache?: boolean; + item_id: number; /** - * @description The FLUX Redux image prompt. - * @default null + * Batch Id + * @description The ID of the queue batch */ - image?: components["schemas"]["ImageField"] | null; + batch_id: string; /** - * @description The bool mask associated with this FLUX Redux image prompt. Excluded regions should be set to False, included regions should be set to True. + * Origin + * @description The origin of the queue item * @default null */ - mask?: components["schemas"]["TensorField"] | null; + origin: string | null; /** - * FLUX Redux Model - * @description The FLUX Redux model to use. + * Destination + * @description The destination of the queue item * @default null */ - redux_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Downsampling Factor - * @description Redux Downsampling Factor (1-9) - * @default 1 - */ - downsampling_factor?: number; - /** - * Downsampling Function - * @description Redux Downsampling Function - * @default area - * @enum {string} - */ - downsampling_function?: "nearest" | "bilinear" | "bicubic" | "area" | "nearest-exact"; + destination: string | null; /** - * Weight - * @description Redux weight (0.0-1.0) - * @default 1 + * User Id + * @description The ID of the user who created the queue item + * @default system */ - weight?: number; + user_id: string; /** - * type - * @default flux_redux_linked - * @constant + * Session Id + * @description The ID of the session (aka graph execution state) */ - type: "flux_redux_linked"; + session_id: string; /** - * FLUX Redux Conditioning List - * @description FLUX Redux conditioning list - * @default null + * Invocation + * @description The ID of the invocation */ - redux_conditioning_list?: components["schemas"]["FluxReduxConditioningField"] | components["schemas"]["FluxReduxConditioningField"][] | null; - }; - /** FluxReduxListOutput */ - FluxReduxListOutput: { + invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; /** - * FLUX Redux Conditioning List - * @description FLUX Redux conditioning tensor + * Invocation Source Id + * @description The ID of the prepared invocation's source node */ - redux_conditioning_list: components["schemas"]["FluxReduxConditioningField"][]; + invocation_source_id: string; /** - * type - * @default flux_redux_list_output - * @constant + * Result + * @description The result of the invocation */ - type: "flux_redux_list_output"; + result: components["schemas"]["BooleanCollectionOutput"] | components["schemas"]["BooleanOutput"] | components["schemas"]["BoundingBoxCollectionOutput"] | components["schemas"]["BoundingBoxOutput"] | components["schemas"]["CLIPOutput"] | components["schemas"]["CLIPSkipInvocationOutput"] | components["schemas"]["CalculateImageTilesOutput"] | components["schemas"]["CogView4ConditioningOutput"] | components["schemas"]["CogView4ModelLoaderOutput"] | components["schemas"]["CollectInvocationOutput"] | components["schemas"]["ColorCollectionOutput"] | components["schemas"]["ColorOutput"] | components["schemas"]["ConditioningCollectionOutput"] | components["schemas"]["ConditioningOutput"] | components["schemas"]["ControlOutput"] | components["schemas"]["DenoiseMaskOutput"] | components["schemas"]["FaceMaskOutput"] | components["schemas"]["FaceOffOutput"] | components["schemas"]["FloatCollectionOutput"] | components["schemas"]["FloatGeneratorOutput"] | components["schemas"]["FloatOutput"] | components["schemas"]["Flux2KleinLoRALoaderOutput"] | components["schemas"]["Flux2KleinModelLoaderOutput"] | components["schemas"]["FluxConditioningCollectionOutput"] | components["schemas"]["FluxConditioningOutput"] | components["schemas"]["FluxControlLoRALoaderOutput"] | components["schemas"]["FluxControlNetOutput"] | components["schemas"]["FluxFillOutput"] | components["schemas"]["FluxKontextOutput"] | components["schemas"]["FluxLoRALoaderOutput"] | components["schemas"]["FluxModelLoaderOutput"] | components["schemas"]["FluxReduxOutput"] | components["schemas"]["GradientMaskOutput"] | components["schemas"]["IPAdapterOutput"] | components["schemas"]["IdealSizeOutput"] | components["schemas"]["ImageCollectionOutput"] | components["schemas"]["ImageGeneratorOutput"] | components["schemas"]["ImageOutput"] | components["schemas"]["ImagePanelCoordinateOutput"] | components["schemas"]["IntegerCollectionOutput"] | components["schemas"]["IntegerGeneratorOutput"] | components["schemas"]["IntegerOutput"] | components["schemas"]["IterateInvocationOutput"] | components["schemas"]["LatentsCollectionOutput"] | components["schemas"]["LatentsMetaOutput"] | components["schemas"]["LatentsOutput"] | components["schemas"]["LoRALoaderOutput"] | components["schemas"]["LoRASelectorOutput"] | components["schemas"]["MDControlListOutput"] | components["schemas"]["MDIPAdapterListOutput"] | components["schemas"]["MDT2IAdapterListOutput"] | components["schemas"]["MaskOutput"] | components["schemas"]["MetadataItemOutput"] | components["schemas"]["MetadataOutput"] | components["schemas"]["MetadataToLorasCollectionOutput"] | components["schemas"]["MetadataToModelOutput"] | components["schemas"]["MetadataToSDXLModelOutput"] | components["schemas"]["ModelIdentifierOutput"] | components["schemas"]["ModelLoaderOutput"] | components["schemas"]["NoiseOutput"] | components["schemas"]["PBRMapsOutput"] | components["schemas"]["PairTileImageOutput"] | components["schemas"]["PromptTemplateOutput"] | components["schemas"]["SD3ConditioningOutput"] | components["schemas"]["SDXLLoRALoaderOutput"] | components["schemas"]["SDXLModelLoaderOutput"] | components["schemas"]["SDXLRefinerModelLoaderOutput"] | components["schemas"]["SchedulerOutput"] | components["schemas"]["Sd3ModelLoaderOutput"] | components["schemas"]["SeamlessModeOutput"] | components["schemas"]["String2Output"] | components["schemas"]["StringCollectionOutput"] | components["schemas"]["StringGeneratorOutput"] | components["schemas"]["StringOutput"] | components["schemas"]["StringPosNegOutput"] | components["schemas"]["T2IAdapterOutput"] | components["schemas"]["TileToPropertiesOutput"] | components["schemas"]["UNetOutput"] | components["schemas"]["VAEOutput"] | components["schemas"]["ZImageConditioningOutput"] | components["schemas"]["ZImageControlOutput"] | components["schemas"]["ZImageLoRALoaderOutput"] | components["schemas"]["ZImageModelLoaderOutput"]; }; /** - * FluxReduxOutput - * @description The conditioning output of a FLUX Redux invocation. + * InvocationErrorEvent + * @description Event model for invocation_error */ - FluxReduxOutput: { - /** - * Conditioning - * @description FLUX Redux conditioning tensor - */ - redux_cond: components["schemas"]["FluxReduxConditioningField"]; + InvocationErrorEvent: { /** - * type - * @default flux_redux_output - * @constant + * Timestamp + * @description The timestamp of the event */ - type: "flux_redux_output"; - }; - /** - * Rescale FLUX Redux Conditioning - * @description Rescales a FLUX Redux conditioning field to a target max norm. - */ - FluxReduxRescaleConditioningInvocation: { + timestamp: number; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Queue Id + * @description The ID of the queue */ - id: string; + queue_id: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Item Id + * @description The ID of the queue item */ - is_intermediate?: boolean; + item_id: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Batch Id + * @description The ID of the queue batch */ - use_cache?: boolean; + batch_id: string; /** - * @description FLUX Redux Conditioning to rescale + * Origin + * @description The origin of the queue item * @default null */ - redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | null; + origin: string | null; /** - * Target Norm - * @description The target max norm for the conditioning. - * @default 1 + * Destination + * @description The destination of the queue item + * @default null */ - target_norm?: number; + destination: string | null; /** - * type - * @default flux_redux_rescale_conditioning - * @constant + * User Id + * @description The ID of the user who created the queue item + * @default system */ - type: "flux_redux_rescale_conditioning"; - }; - /** - * Rescale FLUX Conditioning - * @description Rescales a FLUX conditioning field to a target max norm. - */ - FluxRescaleConditioningInvocation: { + user_id: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Session Id + * @description The ID of the session (aka graph execution state) */ - id: string; + session_id: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Invocation + * @description The ID of the invocation */ - is_intermediate?: boolean; + invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Invocation Source Id + * @description The ID of the prepared invocation's source node */ - use_cache?: boolean; + invocation_source_id: string; /** - * @description FLUX Conditioning to rescale - * @default null + * Error Type + * @description The error type */ - conditioning?: components["schemas"]["FluxConditioningField"] | null; + error_type: string; /** - * Clip Rescale - * @description Whether to rescale the CLIP embeddings. - * @default true + * Error Message + * @description The error message */ - clip_rescale?: boolean; + error_message: string; /** - * Clip Target Norm - * @description The target max norm for the CLIP embeddings. - * @default 30 + * Error Traceback + * @description The error traceback */ - clip_target_norm?: number; - /** - * T5 Rescale - * @description Whether to rescale the T5 embeddings. - * @default true - */ - t5_rescale?: boolean; - /** - * T5 Target Norm - * @description The target max norm for the T5 embeddings. - * @default 10 - */ - t5_target_norm?: number; - /** - * type - * @default flux_rescale_conditioning - * @constant - */ - type: "flux_rescale_conditioning"; - }; - /** - * Scale FLUX Conditioning - * @description Scales a FLUX conditioning field by a factor. - */ - FluxScaleConditioningInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description FLUX Conditioning to scale - * @default null - */ - conditioning?: components["schemas"]["FluxConditioningField"] | null; - /** - * Scale - * @description Scaling factor - * @default 1 - */ - scale?: number; - /** - * Negative - * @description Scale negative conditioning - * @default false - */ - negative?: boolean; - /** - * type - * @default flux_scale_conditioning - * @constant - */ - type: "flux_scale_conditioning"; - }; - /** - * Scale FLUX Prompt Section(s) - * @description Scales one or more sections of a FLUX prompt conditioning. - */ - FluxScalePromptSectionInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description FLUX Conditioning to modify - * @default null - */ - conditioning?: components["schemas"]["FluxConditioningField"] | null; - /** - * T5Encoder - * @description T5 Encoder model and tokenizer used for the original conditioning. - * @default null - */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; - /** - * Prompt - * @description The full prompt text used for the original conditioning. - * @default null - */ - prompt?: string | null; - /** - * Prompt Section - * @description The section or sections of the prompt to scale. - * @default null - */ - prompt_section?: string | string[] | null; - /** - * Scale - * @description The scaling factor or factors for the prompt section(s). - * @default 1 - */ - scale?: number | number[]; - /** - * Positions - * @description The start character position(s) of the section(s) to scale. If provided, this is used to locate the section(s) instead of searching. - * @default null - */ - positions?: number | number[] | null; - /** - * Rescale Output - * @description Rescales the output T5 embeddings to have the same max vector norm as the original conditioning. - * @default false - */ - rescale_output?: boolean; - /** - * type - * @default flux_scale_prompt_section - * @constant - */ - type: "flux_scale_prompt_section"; - }; - /** - * Scale FLUX Redux Conditioning - * @description Scales a FLUX Redux conditioning field by a factor. - */ - FluxScaleReduxConditioningInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description FLUX Redux Conditioning to scale - * @default null - */ - redux_conditioning?: components["schemas"]["FluxReduxConditioningField"] | null; - /** - * Scale - * @description Scaling factor - * @default 1 - */ - scale?: number; - /** - * Negative - * @description Scale negative conditioning - * @default false - */ - negative?: boolean; - /** - * type - * @default flux_scale_redux_conditioning - * @constant - */ - type: "flux_scale_redux_conditioning"; - }; - /** - * Prompt - FLUX - * @description Encodes and preps a prompt for a flux image. - */ - FluxTextEncoderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip?: components["schemas"]["CLIPField"] | null; - /** - * T5Encoder - * @description T5 tokenizer and text encoder - * @default null - */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; - /** - * T5 Max Seq Len - * @description Max sequence length for the T5 encoder. Expected to be 256 for FLUX schnell models and 512 for FLUX dev models. - * @default null - */ - t5_max_seq_len?: (256 | 512) | null; - /** - * Prompt - * @description Text prompt to encode. - * @default null - */ - prompt?: string | null; - /** - * @description A mask defining the region that this conditioning prompt applies to. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default flux_text_encoder - * @constant - */ - type: "flux_text_encoder"; - }; - /** - * Prompt - FLUX Linked - * @description Collects FLUX prompt conditionings to pass to other nodes. - */ - FluxTextEncoderLinkedInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip?: components["schemas"]["CLIPField"] | null; - /** - * T5Encoder - * @description T5 tokenizer and text encoder - * @default null - */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; - /** - * T5 Max Seq Len - * @description Max sequence length for the T5 encoder. Expected to be 256 for FLUX schnell models and 512 for FLUX dev models. - * @default null - */ - t5_max_seq_len?: (256 | 512) | null; - /** - * Prompt - * @description Text prompt to encode. - * @default null - */ - prompt?: string | null; - /** - * @description A mask defining the region that this conditioning prompt applies to. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default flux_text_encoder_linked - * @constant - */ - type: "flux_text_encoder_linked"; - /** - * FLUX Text Encoder Conditioning List - * @description Conditioning tensor - * @default null - */ - flux_text_encoder_list?: components["schemas"]["FluxConditioningField"] | components["schemas"]["FluxConditioningField"][] | null; - }; - /** - * FluxTextEncoderListOutput - * @description Output for a list of FLUX text encoder conditioning. - */ - FluxTextEncoderListOutput: { - /** - * FLUX Text Encoder Conditioning List - * @description Conditioning tensor - */ - flux_text_encoder_list: components["schemas"]["FluxConditioningField"][]; - /** - * type - * @default flux_text_encoder_list_output - * @constant - */ - type: "flux_text_encoder_list_output"; - }; - /** - * Latents to Image - FLUX - * @description Generates an image from latents. - */ - FluxVaeDecodeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * @description VAE - * @default null - */ - vae?: components["schemas"]["VAEField"] | null; - /** - * type - * @default flux_vae_decode - * @constant - */ - type: "flux_vae_decode"; - }; - /** - * Image to Latents - FLUX - * @description Encodes an image into latents. - */ - FluxVaeEncodeInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to encode. - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description VAE - * @default null - */ - vae?: components["schemas"]["VAEField"] | null; - /** - * type - * @default flux_vae_encode - * @constant - */ - type: "flux_vae_encode"; - }; - /** - * FluxVariantType - * @description FLUX.1 model variants. - * @enum {string} - */ - FluxVariantType: "schnell" | "dev" | "dev_fill"; - /** - * FLUX Weighted Prompt - * @description Parses a weighted prompt, then encodes it for FLUX. - */ - FluxWeightedPromptInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Prompt - * @description Text prompt to encode. - * @default null - */ - prompt?: string | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip?: components["schemas"]["CLIPField"] | null; - /** - * T5Encoder - * @description T5 tokenizer and text encoder - * @default null - */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; - /** - * T5 Max Seq Len - * @description Max sequence length for the T5 encoder. Expected to be 256 for FLUX schnell models and 512 for FLUX dev models. - * @default null - */ - t5_max_seq_len?: (256 | 512) | null; - /** - * @description A mask defining the region that this conditioning prompt applies to. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * Rescale Output - * @description Rescales the output T5 embeddings to have the same max vector norm as the original conditioning. - * @default false - */ - rescale_output?: boolean; - /** - * type - * @default flux_weighted_prompt - * @constant - */ - type: "flux_weighted_prompt"; - }; - /** - * FluxWeightedPromptOutput - * @description Outputs a FLUX conditioning and a cleaned prompt - */ - FluxWeightedPromptOutput: { - /** @description The FLUX conditioning */ - conditioning: components["schemas"]["FluxConditioningField"]; - /** - * Cleaned Prompt - * @description The prompt with all weighting syntax removed - */ - cleaned_prompt: string; - /** - * type - * @default flux_weighted_prompt_output - * @constant - */ - type: "flux_weighted_prompt_output"; - }; - /** FoundModel */ - FoundModel: { - /** - * Path - * @description Path to the model - */ - path: string; - /** - * Is Installed - * @description Whether or not the model is already installed - */ - is_installed: boolean; - }; - /** - * FreeUConfig - * @description Configuration for the FreeU hyperparameters. - * - https://huggingface.co/docs/diffusers/main/en/using-diffusers/freeu - * - https://github.com/ChenyangSi/FreeU - */ - FreeUConfig: { - /** - * S1 - * @description Scaling factor for stage 1 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. - */ - s1: number; - /** - * S2 - * @description Scaling factor for stage 2 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. - */ - s2: number; - /** - * B1 - * @description Scaling factor for stage 1 to amplify the contributions of backbone features. - */ - b1: number; - /** - * B2 - * @description Scaling factor for stage 2 to amplify the contributions of backbone features. - */ - b2: number; - }; - /** - * Apply FreeU - SD1.5, SDXL - * @description Applies FreeU to the UNet. Suggested values (b1/b2/s1/s2): - * - * SD1.5: 1.2/1.4/0.9/0.2, - * SD2: 1.1/1.2/0.9/0.2, - * SDXL: 1.1/1.2/0.6/0.4, - */ - FreeUInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; - /** - * B1 - * @description Scaling factor for stage 1 to amplify the contributions of backbone features. - * @default 1.2 - */ - b1?: number; - /** - * B2 - * @description Scaling factor for stage 2 to amplify the contributions of backbone features. - * @default 1.4 - */ - b2?: number; - /** - * S1 - * @description Scaling factor for stage 1 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. - * @default 0.9 - */ - s1?: number; - /** - * S2 - * @description Scaling factor for stage 2 to attenuate the contributions of the skip features. This is done to mitigate the "oversmoothing effect" in the enhanced denoising process. - * @default 0.2 - */ - s2?: number; - /** - * type - * @default freeu - * @constant - */ - type: "freeu"; - }; - /** Frequency Blend Latents */ - FrequencyBlendLatents: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Image from which to take high frequencies. - * @default null - */ - high_input?: components["schemas"]["LatentsField"] | null; - /** - * @description Image from which to take low frequencies. - * @default null - */ - low_input?: components["schemas"]["LatentsField"] | null; - /** - * Threshold - * @description How much of the high-frequency to use. Negative values swap the inputs. - * @default 0.05 - */ - threshold?: number; - /** - * Weighted Blend - * @description Weighted blend? - * @default true - */ - weighted_blend?: boolean; - /** - * type - * @default frequency_blend_latents - * @constant - */ - type: "frequency_blend_latents"; - }; - /** - * Frequency Spectrum Match Latents - * @description Generates latent noise with the frequency spectrum of target latents, masked by a provided mask image. - * Takes both target latents and white noise latents as inputs. - */ - FrequencySpectrumMatchLatentsInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Target latents to match frequency spectrum (A) - * @default null - */ - target_latents_a?: components["schemas"]["LatentsField"] | null; - /** - * @description Target latents to match frequency spectrum (B) (optional) - * @default null - */ - target_latents_b?: components["schemas"]["LatentsField"] | null; - /** - * Frequency Blend Alpha - * @description Blend ratio for the frequency spectra - * @default 0 - */ - frequency_blend_alpha?: number; - /** - * @description White noise latents for phase information (A) - * @default null - */ - phase_latents_a?: components["schemas"]["LatentsField"] | null; - /** - * @description White noise latents for phase information (B) (optional) - * @default null - */ - phase_latents_b?: components["schemas"]["LatentsField"] | null; - /** - * Phase Blend Alpha - * @description Blend ratio for the phases - * @default 0 - */ - phase_blend_alpha?: number; - /** - * @description Mask for blending (optional) - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Blur Sigma - * @description Amount of Gaussian blur to apply to the mask - * @default 0 - */ - blur_sigma?: number; - /** - * type - * @default frequency_match_latents - * @constant - */ - type: "frequency_match_latents"; - }; - /** - * Generate Evolutionary Prompts - * @description Generates a new population of prompts by performing crossover operations - * on an input list of parent seed vectors. - */ - GenerateEvolutionaryPromptsInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Resolutions Dict - * @description Private field for id substitutions dict cache - * @default {} - */ - resolutions_dict?: { - [key: string]: unknown; - }; - /** - * Lookups - * @description Lookup table(s) containing template(s) (JSON) - * @default [] - */ - lookups?: string | string[]; - /** - * Remove Negatives - * @description Whether to strip out text between [] - * @default false - */ - remove_negatives?: boolean; - /** - * Strip Parens Probability - * @description Probability of removing attention group weightings - * @default 0 - */ - strip_parens_probability?: number; - /** - * Resolutions - * @description JSON structure of substitutions by id by tag - * @default [] - */ - resolutions?: string | string[]; - /** - * Seed Vectors In - * @description List of JSON array strings, each representing a parent's seed vector. - * @default [] - */ - seed_vectors_in?: string[]; - /** - * Target Population Size - * @description The desired size of the new population to generate. - * @default 10 - */ - target_population_size?: number; - /** - * Selected Pair - * @description The selected population member to output specifically. - * @default 0 - */ - selected_pair?: number; - /** - * Ga Seed - * @description Seed for the random number generator to ensure deterministic GA operations - * @default 0 - */ - ga_seed?: number; - /** - * Crossover Non Terminal - * @description Optional: The non-terminal (key in lookups) to target for the crossover branch. If None, a random one will be chosen from available common non-terminals for each crossover. - * @default null - */ - crossover_non_terminal?: string | null; - /** - * type - * @default generate_evolutionary_prompts - * @constant - */ - type: "generate_evolutionary_prompts"; - }; - /** - * GeneratePasswordResponse - * @description Response containing a generated password. - */ - GeneratePasswordResponse: { - /** - * Password - * @description Generated strong password - */ - password: string; - }; - /** - * Get Image Mask Bounding Box - * @description Gets the bounding box of the given mask image. - */ - GetMaskBoundingBoxInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The mask to crop. - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Margin - * @description Margin to add to the bounding box. - * @default 0 - */ - margin?: number; - /** - * @description Color of the mask in the image. - * @default { - * "r": 255, - * "g": 255, - * "b": 255, - * "a": 255 - * } - */ - mask_color?: components["schemas"]["ColorField"]; - /** - * type - * @default get_image_mask_bounding_box - * @constant - */ - type: "get_image_mask_bounding_box"; - }; - /** - * Get Source Frame Rate - * @description Get the source framerate of an MP4 video and provide it as output. - */ - GetSourceFrameRateInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Video Path - * @description The path to the MP4 video file - * @default null - */ - video_path?: string | null; - /** - * type - * @default get_source_framerate - * @constant - */ - type: "get_source_framerate"; - }; - /** - * Get Total Frames - * @description Get the total number of frames in an MP4 video and provide it as output. - */ - GetTotalFramesInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Video Path - * @description The path to the MP4 video file - * @default null - */ - video_path?: string | null; - /** - * type - * @default get_total_frames - * @constant - */ - type: "get_total_frames"; - }; - /** GlmEncoderField */ - GlmEncoderField: { - /** @description Info to load tokenizer submodel */ - tokenizer: components["schemas"]["ModelIdentifierField"]; - /** @description Info to load text_encoder submodel */ - text_encoder: components["schemas"]["ModelIdentifierField"]; - }; - /** - * GradientMaskOutput - * @description Outputs a denoise mask and an image representing the total gradient of the mask. - */ - GradientMaskOutput: { - /** @description Mask for denoise model run. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. */ - denoise_mask: components["schemas"]["DenoiseMaskField"]; - /** @description Image representing the total gradient area of the mask. For paste-back purposes. */ - expanded_mask_area: components["schemas"]["ImageField"]; - /** - * type - * @default gradient_mask_output - * @constant - */ - type: "gradient_mask_output"; - }; - /** Graph */ - Graph: { - /** - * Id - * @description The id of this graph - */ - id?: string; - /** - * Nodes - * @description The nodes in this graph - */ - nodes?: { - [key: string]: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; - }; - /** - * Edges - * @description The connections between nodes and their fields in this graph - */ - edges?: components["schemas"]["Edge"][]; - }; - /** - * GraphExecutionState - * @description Tracks the state of a graph execution - */ - GraphExecutionState: { - /** - * Id - * @description The id of the execution state - */ - id: string; - /** @description The graph being executed */ - graph: components["schemas"]["Graph"]; - /** @description The expanded graph of activated and executed nodes */ - execution_graph: components["schemas"]["Graph"]; - /** - * Executed - * @description The set of node ids that have been executed - */ - executed: string[]; - /** - * Executed History - * @description The list of node ids that have been executed, in order of execution - */ - executed_history: string[]; - /** - * Results - * @description The results of node executions - */ - results: { - [key: string]: components["schemas"]["BooleanCollectionOutput"] | components["schemas"]["BooleanOutput"] | components["schemas"]["BoundingBoxCollectionOutput"] | components["schemas"]["BoundingBoxOutput"] | components["schemas"]["CLIPOutput"] | components["schemas"]["CLIPSkipInvocationOutput"] | components["schemas"]["CMYKSeparationOutput"] | components["schemas"]["CMYKSplitOutput"] | components["schemas"]["CalculateImageTilesOutput"] | components["schemas"]["ClipsegMaskHierarchyOutput"] | components["schemas"]["CogView4ConditioningOutput"] | components["schemas"]["CogView4ModelLoaderOutput"] | components["schemas"]["CollectInvocationOutput"] | components["schemas"]["CollectionCountOutput"] | components["schemas"]["CollectionIndexOutput"] | components["schemas"]["CollectionJoinOutput"] | components["schemas"]["CollectionReverseOutput"] | components["schemas"]["CollectionSliceOutput"] | components["schemas"]["CollectionSortOutput"] | components["schemas"]["CollectionUniqueOutput"] | components["schemas"]["ColorCollectionOutput"] | components["schemas"]["ColorOutput"] | components["schemas"]["ConditioningCollectionOutput"] | components["schemas"]["ConditioningOutput"] | components["schemas"]["ControlListOutput"] | components["schemas"]["ControlOutput"] | components["schemas"]["DenoiseMaskOutput"] | components["schemas"]["EscapedOutput"] | components["schemas"]["EvolutionaryPromptListOutput"] | components["schemas"]["FaceMaskOutput"] | components["schemas"]["FaceOffOutput"] | components["schemas"]["FloatCollectionOutput"] | components["schemas"]["FloatGeneratorOutput"] | components["schemas"]["FloatOutput"] | components["schemas"]["Flux2KleinLoRALoaderOutput"] | components["schemas"]["Flux2KleinModelLoaderOutput"] | components["schemas"]["FluxConditioningBlendOutput"] | components["schemas"]["FluxConditioningCollectionOutput"] | components["schemas"]["FluxConditioningDeltaAndAugmentedOutput"] | components["schemas"]["FluxConditioningListOutput"] | components["schemas"]["FluxConditioningOutput"] | components["schemas"]["FluxConditioningStoreOutput"] | components["schemas"]["FluxControlLoRALoaderOutput"] | components["schemas"]["FluxControlNetCollectionOutput"] | components["schemas"]["FluxControlNetListOutput"] | components["schemas"]["FluxControlNetOutput"] | components["schemas"]["FluxFillOutput"] | components["schemas"]["FluxIdealSizeOutput"] | components["schemas"]["FluxKontextOutput"] | components["schemas"]["FluxLoRALoaderOutput"] | components["schemas"]["FluxModelLoaderOutput"] | components["schemas"]["FluxReduxCollectionOutput"] | components["schemas"]["FluxReduxListOutput"] | components["schemas"]["FluxReduxOutput"] | components["schemas"]["FluxTextEncoderListOutput"] | components["schemas"]["FluxWeightedPromptOutput"] | components["schemas"]["GradientMaskOutput"] | components["schemas"]["HalvedPromptOutput"] | components["schemas"]["IPAdapterListOutput"] | components["schemas"]["IPAdapterOutput"] | components["schemas"]["IdealSizeOutput"] | components["schemas"]["ImageCollectionOutput"] | components["schemas"]["ImageGeneratorOutput"] | components["schemas"]["ImageIndexCollectOutput"] | components["schemas"]["ImageOutput"] | components["schemas"]["ImagePanelCoordinateOutput"] | components["schemas"]["ImageSOMOutput"] | components["schemas"]["ImageToImageNameOutput"] | components["schemas"]["ImageToXYImageTilesOutput"] | components["schemas"]["ImagesIndexToVideoOutput"] | components["schemas"]["IntegerCollectionOutput"] | components["schemas"]["IntegerGeneratorOutput"] | components["schemas"]["IntegerOutput"] | components["schemas"]["IterateInvocationOutput"] | components["schemas"]["JsonListStringsOutput"] | components["schemas"]["LatentsCollectionOutput"] | components["schemas"]["LatentsMetaOutput"] | components["schemas"]["LatentsOutput"] | components["schemas"]["LoRACollectionFromPathOutput"] | components["schemas"]["LoRACollectionOutput"] | components["schemas"]["LoRACollectionToggleOutput"] | components["schemas"]["LoRALoaderOutput"] | components["schemas"]["LoRANameGrabberOutput"] | components["schemas"]["LoRASelectorOutput"] | components["schemas"]["LoRAToggleOutput"] | components["schemas"]["LoadAllTextFilesInFolderOutput"] | components["schemas"]["LoadTextFileToStringOutput"] | components["schemas"]["LookupLoRACollectionTriggersOutput"] | components["schemas"]["LookupLoRATriggersOutput"] | components["schemas"]["LookupTableOutput"] | components["schemas"]["MDControlListOutput"] | components["schemas"]["MDIPAdapterListOutput"] | components["schemas"]["MDT2IAdapterListOutput"] | components["schemas"]["MaskOutput"] | components["schemas"]["MathEvalOutput"] | components["schemas"]["MergeLoRACollectionsOutput"] | components["schemas"]["MergeStringCollectionsOutput"] | components["schemas"]["MetadataItemOutput"] | components["schemas"]["MetadataOutput"] | components["schemas"]["MetadataToLorasCollectionOutput"] | components["schemas"]["MetadataToModelOutput"] | components["schemas"]["MetadataToSDXLModelOutput"] | components["schemas"]["ModelIdentifierOutput"] | components["schemas"]["ModelLoaderOutput"] | components["schemas"]["ModelNameGrabberOutput"] | components["schemas"]["ModelToggleOutput"] | components["schemas"]["NightmareOutput"] | components["schemas"]["NoiseOutput"] | components["schemas"]["PBRMapsOutput"] | components["schemas"]["PTFieldsCollectOutput"] | components["schemas"]["PTFieldsExpandOutput"] | components["schemas"]["PairTileImageOutput"] | components["schemas"]["PaletteOutput"] | components["schemas"]["PrintStringToConsoleOutput"] | components["schemas"]["PromptStrengthOutput"] | components["schemas"]["PromptTemplateOutput"] | components["schemas"]["PromptsToFileInvocationOutput"] | components["schemas"]["PrunedPromptOutput"] | components["schemas"]["Qwen3PromptProOutput"] | components["schemas"]["RGBSplitOutput"] | components["schemas"]["RandomImageSizeOutput"] | components["schemas"]["RandomLoRAMixerOutput"] | components["schemas"]["ReapplyLoRAWeightOutput"] | components["schemas"]["RetrieveFluxConditioningMultiOutput"] | components["schemas"]["SD3ConditioningOutput"] | components["schemas"]["SDXLLoRALoaderOutput"] | components["schemas"]["SDXLModelLoaderOutput"] | components["schemas"]["SDXLRefinerModelLoaderOutput"] | components["schemas"]["SchedulerOutput"] | components["schemas"]["Sd3ModelLoaderOutput"] | components["schemas"]["SeamlessModeOutput"] | components["schemas"]["ShadowsHighlightsMidtonesMasksOutput"] | components["schemas"]["String2Output"] | components["schemas"]["StringCollectionJoinerOutput"] | components["schemas"]["StringCollectionOutput"] | components["schemas"]["StringGeneratorOutput"] | components["schemas"]["StringOutput"] | components["schemas"]["StringPosNegOutput"] | components["schemas"]["StringToLoraOutput"] | components["schemas"]["StringToMainModelOutput"] | components["schemas"]["StringToModelOutput"] | components["schemas"]["StringToSDXLModelOutput"] | components["schemas"]["T2IAdapterListOutput"] | components["schemas"]["T2IAdapterOutput"] | components["schemas"]["ThresholdingOutput"] | components["schemas"]["TileSizeOutput"] | components["schemas"]["TileToPropertiesOutput"] | components["schemas"]["TilesOutput"] | components["schemas"]["TraceryOutput"] | components["schemas"]["UNetOutput"] | components["schemas"]["VAEOutput"] | components["schemas"]["WeightedStringOutput"] | components["schemas"]["XYExpandOutput"] | components["schemas"]["XYImageExpandOutput"] | components["schemas"]["XYProductOutput"] | components["schemas"]["ZImageConditioningOutput"] | components["schemas"]["ZImageControlOutput"] | components["schemas"]["ZImageLoRALoaderOutput"] | components["schemas"]["ZImageModelLoaderOutput"]; - }; - /** - * Errors - * @description Errors raised when executing nodes - */ - errors: { - [key: string]: string; - }; - /** - * Prepared Source Mapping - * @description The map of prepared nodes to original graph nodes - */ - prepared_source_mapping: { - [key: string]: string; - }; - /** - * Source Prepared Mapping - * @description The map of original graph nodes to prepared nodes - */ - source_prepared_mapping: { - [key: string]: string[]; - }; - /** Ready Order */ - ready_order?: string[]; - /** - * Indegree - * @description Remaining unmet input count for exec nodes - */ - indegree?: { - [key: string]: number; - }; - }; - /** - * Grounding DINO (Text Prompt Object Detection) - * @description Runs a Grounding DINO model. Performs zero-shot bounding-box object detection from a text prompt. - */ - GroundingDinoInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Model - * @description The Grounding DINO model to use. - * @default null - */ - model?: ("grounding-dino-tiny" | "grounding-dino-base") | null; - /** - * Prompt - * @description The prompt describing the object to segment. - * @default null - */ - prompt?: string | null; - /** - * @description The image to segment. - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Detection Threshold - * @description The detection threshold for the Grounding DINO model. All detected bounding boxes with scores above this threshold will be returned. - * @default 0.3 - */ - detection_threshold?: number; - /** - * type - * @default grounding_dino - * @constant - */ - type: "grounding_dino"; - }; - /** - * HED Edge Detection - * @description Geneartes an edge map using the HED (softedge) model. - */ - HEDEdgeDetectionInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Scribble - * @description Whether or not to use scribble mode - * @default false - */ - scribble?: boolean; - /** - * type - * @default hed_edge_detection - * @constant - */ - type: "hed_edge_detection"; - }; - /** - * HFModelSource - * @description A HuggingFace repo_id with optional variant, sub-folder(s) and access token. - * Note that the variant option, if not provided to the constructor, will default to fp16, which is - * what people (almost) always want. - * - * The subfolder can be a single path or multiple paths joined by '+' (e.g., "text_encoder+tokenizer"). - * When multiple subfolders are specified, all of them will be downloaded and combined into the model directory. - */ - HFModelSource: { - /** Repo Id */ - repo_id: string; - /** @default fp16 */ - variant?: components["schemas"]["ModelRepoVariant"] | null; - /** Subfolder */ - subfolder?: string | null; - /** Access Token */ - access_token?: string | null; - /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} - */ - type: "hf"; - }; - /** - * HFTokenStatus - * @enum {string} - */ - HFTokenStatus: "valid" | "invalid" | "unknown"; - /** HTTPValidationError */ - HTTPValidationError: { - /** Detail */ - detail?: components["schemas"]["ValidationError"][]; - }; - /** - * Halftone - * @description Halftones an image - */ - HalftoneInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to halftone - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Spacing - * @description Halftone dot spacing - * @default 8 - */ - spacing?: number; - /** - * Angle - * @description Halftone angle - * @default 45 - */ - angle?: number; - /** - * Oversampling - * @description Oversampling factor - * @default 1 - */ - oversampling?: number; - /** - * type - * @default halftone - * @constant - */ - type: "halftone"; - }; - /** - * HalvedPromptOutput - * @description Output for a halved prompt. - */ - HalvedPromptOutput: { - /** - * Prompt - * @description The entire output prompt - * @default - */ - prompt: string; - /** - * Part A - * @description The first part of the output prompt - * @default - */ - part_a: string; - /** - * Part B - * @description The second part of the output prompt - * @default - */ - part_b: string; - /** - * Resolutions - * @description JSON dict of [tagname,id] resolutions - * @default - */ - resolutions: string; - /** - * Seed Vector - * @description JSON string of the seed vector used for generation - * @default - */ - seed_vector: string; - /** - * type - * @default halved_prompt_output - * @constant - */ - type: "halved_prompt_output"; - }; - /** - * Heuristic Resize - * @description Resize an image using a heuristic method. Preserves edge maps. - */ - HeuristicResizeInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to resize - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Width - * @description The width to resize to (px) - * @default 512 - */ - width?: number; - /** - * Height - * @description The height to resize to (px) - * @default 512 - */ - height?: number; - /** - * type - * @default heuristic_resize - * @constant - */ - type: "heuristic_resize"; - }; - /** - * HuggingFaceMetadata - * @description Extended metadata fields provided by HuggingFace. - */ - HuggingFaceMetadata: { - /** - * Name - * @description model's name - */ - name: string; - /** - * Files - * @description model files and their sizes - */ - files?: components["schemas"]["RemoteModelFile"][]; - /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} - */ - type: "huggingface"; - /** - * Id - * @description The HF model id - */ - id: string; - /** - * Api Response - * @description Response from the HF API as stringified JSON - */ - api_response?: string | null; - /** - * Is Diffusers - * @description Whether the metadata is for a Diffusers format model - * @default false - */ - is_diffusers?: boolean; - /** - * Ckpt Urls - * @description URLs for all checkpoint format models in the metadata - */ - ckpt_urls?: string[] | null; - }; - /** HuggingFaceModels */ - HuggingFaceModels: { - /** - * Urls - * @description URLs for all checkpoint format models in the metadata - */ - urls: string[] | null; - /** - * Is Diffusers - * @description Whether the metadata is for a Diffusers format model - */ - is_diffusers: boolean; - }; - /** IPAdapterField */ - IPAdapterField: { - /** - * Image - * @description The IP-Adapter image prompt(s). - */ - image: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][]; - /** @description The IP-Adapter model to use. */ - ip_adapter_model: components["schemas"]["ModelIdentifierField"]; - /** @description The name of the CLIP image encoder model. */ - image_encoder_model: components["schemas"]["ModelIdentifierField"]; - /** - * Weight - * @description The weight given to the IP-Adapter. - * @default 1 - */ - weight?: number | number[]; - /** - * Target Blocks - * @description The IP Adapter blocks to apply - * @default [] - */ - target_blocks?: string[]; - /** - * Method - * @description Weight apply method - * @default full - */ - method?: string; - /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) - * @default 0 - */ - begin_step_percent?: number; - /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) - * @default 1 - */ - end_step_percent?: number; - /** - * @description The bool mask associated with this IP-Adapter. Excluded regions should be set to False, included regions should be set to True. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - }; - /** - * IP-Adapter - SD1.5, SDXL - * @description Collects IP-Adapter info to pass to other nodes. - */ - IPAdapterInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Image - * @description The IP-Adapter image prompt(s). - * @default null - */ - image?: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][] | null; - /** - * IP-Adapter Model - * @description The IP-Adapter model. - * @default null - */ - ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Clip Vision Model - * @description CLIP Vision model to use. Overrides model settings. Mandatory for checkpoint models. - * @default ViT-H - * @enum {string} - */ - clip_vision_model?: "ViT-H" | "ViT-G" | "ViT-L"; - /** - * Weight - * @description The weight given to the IP-Adapter - * @default 1 - */ - weight?: number | number[]; - /** - * Method - * @description The method to apply the IP-Adapter - * @default full - * @enum {string} - */ - method?: "full" | "style" | "composition" | "style_strong" | "style_precise"; - /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) - * @default 0 - */ - begin_step_percent?: number; - /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) - * @default 1 - */ - end_step_percent?: number; - /** - * @description A mask defining the region that this IP-Adapter applies to. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - }; - /** - * IP-Adapter-Linked - * @description Collects IP-Adapter info to pass to other nodes. - */ - IPAdapterLinkedInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Image - * @description The IP-Adapter image prompt(s). - * @default null - */ - image?: components["schemas"]["ImageField"] | components["schemas"]["ImageField"][] | null; - /** - * IP-Adapter Model - * @description The IP-Adapter model. - * @default null - */ - ip_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Clip Vision Model - * @description CLIP Vision model to use. Overrides model settings. Mandatory for checkpoint models. - * @default ViT-H - * @enum {string} - */ - clip_vision_model?: "ViT-H" | "ViT-G" | "ViT-L"; - /** - * Weight - * @description The weight given to the IP-Adapter - * @default 1 - */ - weight?: number | number[]; - /** - * Method - * @description The method to apply the IP-Adapter - * @default full - * @enum {string} - */ - method?: "full" | "style" | "composition" | "style_strong" | "style_precise"; - /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) - * @default 0 - */ - begin_step_percent?: number; - /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) - * @default 1 - */ - end_step_percent?: number; - /** - * @description A mask defining the region that this IP-Adapter applies to. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default ip_adapter_linked - * @constant - */ - type: "ip_adapter_linked"; - /** - * IP-Adapter List - * @description IP-Adapter to apply - * @default null - */ - ip_adapter_list?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; - }; - /** IPAdapterListOutput */ - IPAdapterListOutput: { - /** - * IP-Adapter List - * @description IP-Adapter to apply - */ - ip_adapter_list: components["schemas"]["IPAdapterField"][]; - /** - * type - * @default ip_adapter_list_output - * @constant - */ - type: "ip_adapter_list_output"; - }; - /** - * IPAdapterMetadataField - * @description IP Adapter Field, minus the CLIP Vision Encoder model - */ - IPAdapterMetadataField: { - /** @description The IP-Adapter image prompt. */ - image: components["schemas"]["ImageField"]; - /** @description The IP-Adapter model. */ - ip_adapter_model: components["schemas"]["ModelIdentifierField"]; - /** - * Clip Vision Model - * @description The CLIP Vision model - * @enum {string} - */ - clip_vision_model: "ViT-L" | "ViT-H" | "ViT-G"; - /** - * Method - * @description Method to apply IP Weights with - * @enum {string} - */ - method: "full" | "style" | "composition" | "style_strong" | "style_precise"; - /** - * Weight - * @description The weight given to the IP-Adapter - */ - weight: number | number[]; - /** - * Begin Step Percent - * @description When the IP-Adapter is first applied (% of total steps) - */ - begin_step_percent: number; - /** - * End Step Percent - * @description When the IP-Adapter is last applied (% of total steps) - */ - end_step_percent: number; - }; - /** IPAdapterOutput */ - IPAdapterOutput: { - /** - * IP-Adapter - * @description IP-Adapter to apply - */ - ip_adapter: components["schemas"]["IPAdapterField"]; - /** - * type - * @default ip_adapter_output - * @constant - */ - type: "ip_adapter_output"; - }; - /** - * IPAdapterRecallParameter - * @description IP Adapter configuration for recall - */ - IPAdapterRecallParameter: { - /** - * Model Name - * @description The name of the IP Adapter model - */ - model_name: string; - /** - * Image Name - * @description The filename of the reference image in outputs/images - */ - image_name?: string | null; - /** - * Weight - * @description The weight for the IP Adapter - * @default 1 - */ - weight?: number; - /** - * Begin Step Percent - * @description When the IP Adapter is first applied (% of total steps) - */ - begin_step_percent?: number | null; - /** - * End Step Percent - * @description When the IP Adapter is last applied (% of total steps) - */ - end_step_percent?: number | null; - /** - * Method - * @description The IP Adapter method - */ - method?: ("full" | "style" | "composition") | null; - /** - * Image Influence - * @description FLUX Redux image influence (if model is flux_redux) - */ - image_influence?: ("lowest" | "low" | "medium" | "high" | "highest") | null; - }; - /** IPAdapter_Checkpoint_FLUX_Config */ - IPAdapter_Checkpoint_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - /** - * Base - * @default flux - * @constant - */ - base: "flux"; - }; - /** IPAdapter_Checkpoint_SD1_Config */ - IPAdapter_Checkpoint_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - /** - * Base - * @default sd-1 - * @constant - */ - base: "sd-1"; - }; - /** IPAdapter_Checkpoint_SD2_Config */ - IPAdapter_Checkpoint_SD2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - /** - * Base - * @default sd-2 - * @constant - */ - base: "sd-2"; - }; - /** IPAdapter_Checkpoint_SDXL_Config */ - IPAdapter_Checkpoint_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - /** - * Base - * @default sdxl - * @constant - */ - base: "sdxl"; - }; - /** IPAdapter_InvokeAI_SD1_Config */ - IPAdapter_InvokeAI_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - /** - * Format - * @default invokeai - * @constant - */ - format: "invokeai"; - /** Image Encoder Model Id */ - image_encoder_model_id: string; - /** - * Base - * @default sd-1 - * @constant - */ - base: "sd-1"; - }; - /** IPAdapter_InvokeAI_SD2_Config */ - IPAdapter_InvokeAI_SD2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - /** - * Format - * @default invokeai - * @constant - */ - format: "invokeai"; - /** Image Encoder Model Id */ - image_encoder_model_id: string; - /** - * Base - * @default sd-2 - * @constant - */ - base: "sd-2"; - }; - /** IPAdapter_InvokeAI_SDXL_Config */ - IPAdapter_InvokeAI_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default ip_adapter - * @constant - */ - type: "ip_adapter"; - /** - * Format - * @default invokeai - * @constant - */ - format: "invokeai"; - /** Image Encoder Model Id */ - image_encoder_model_id: string; - /** - * Base - * @default sdxl - * @constant - */ - base: "sdxl"; - }; - /** - * Ideal Size - SD1.5, SDXL - * @description Calculates the ideal size for generation to avoid duplication - */ - IdealSizeInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Width - * @description Final image width - * @default 1024 - */ - width?: number; - /** - * Height - * @description Final image height - * @default 576 - */ - height?: number; - /** - * @description UNet (scheduler, LoRAs) - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; - /** - * Multiplier - * @description Amount to multiply the model's dimensions by when calculating the ideal size (may result in initial generation artifacts if too large) - * @default 1 - */ - multiplier?: number; - /** - * type - * @default ideal_size - * @constant - */ - type: "ideal_size"; - }; - /** - * IdealSizeOutput - * @description Base class for invocations that output an image - */ - IdealSizeOutput: { - /** - * Width - * @description The ideal width of the image (in pixels) - */ - width: number; - /** - * Height - * @description The ideal height of the image (in pixels) - */ - height: number; - /** - * type - * @default ideal_size_output - * @constant - */ - type: "ideal_size_output"; - }; - /** - * Image Batch - * @description Create a batched generation, where the workflow is executed once for each image in the batch. - */ - ImageBatchInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Batch Group - * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. - * @default None - * @enum {string} - */ - batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; - /** - * Images - * @description The images to batch over - * @default null - */ - images?: components["schemas"]["ImageField"][] | null; - /** - * type - * @default image_batch - * @constant - */ - type: "image_batch"; - }; - /** - * Blur Image - * @description Blurs an image - */ - ImageBlurInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to blur - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Radius - * @description The blur radius - * @default 8 - */ - radius?: number; - /** - * Blur Type - * @description The type of blur - * @default gaussian - * @enum {string} - */ - blur_type?: "gaussian" | "box"; - /** - * type - * @default img_blur - * @constant - */ - type: "img_blur"; - }; - /** - * ImageCategory - * @description The category of an image. - * - * - GENERAL: The image is an output, init image, or otherwise an image without a specialized purpose. - * - MASK: The image is a mask image. - * - CONTROL: The image is a ControlNet control image. - * - USER: The image is a user-provide image. - * - OTHER: The image is some other type of image with a specialized purpose. To be used by external nodes. - * @enum {string} - */ - ImageCategory: "general" | "mask" | "control" | "user" | "other"; - /** - * Extract Image Channel - * @description Gets a channel from an image. - */ - ImageChannelInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to get the channel from - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Channel - * @description The channel to get - * @default A - * @enum {string} - */ - channel?: "A" | "R" | "G" | "B"; - /** - * type - * @default img_chan - * @constant - */ - type: "img_chan"; - }; - /** - * Multiply Image Channel - * @description Scale a specific color channel of an image. - */ - ImageChannelMultiplyInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to adjust - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Channel - * @description Which channel to adjust - * @default null - */ - channel?: ("Red (RGBA)" | "Green (RGBA)" | "Blue (RGBA)" | "Alpha (RGBA)" | "Cyan (CMYK)" | "Magenta (CMYK)" | "Yellow (CMYK)" | "Black (CMYK)" | "Hue (HSV)" | "Saturation (HSV)" | "Value (HSV)" | "Luminosity (LAB)" | "A (LAB)" | "B (LAB)" | "Y (YCbCr)" | "Cb (YCbCr)" | "Cr (YCbCr)") | null; - /** - * Scale - * @description The amount to scale the channel by. - * @default 1 - */ - scale?: number; - /** - * Invert Channel - * @description Invert the channel after scaling - * @default false - */ - invert_channel?: boolean; - /** - * type - * @default img_channel_multiply - * @constant - */ - type: "img_channel_multiply"; - }; - /** - * Offset Image Channel - * @description Add or subtract a value from a specific color channel of an image. - */ - ImageChannelOffsetInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to adjust - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Channel - * @description Which channel to adjust - * @default null - */ - channel?: ("Red (RGBA)" | "Green (RGBA)" | "Blue (RGBA)" | "Alpha (RGBA)" | "Cyan (CMYK)" | "Magenta (CMYK)" | "Yellow (CMYK)" | "Black (CMYK)" | "Hue (HSV)" | "Saturation (HSV)" | "Value (HSV)" | "Luminosity (LAB)" | "A (LAB)" | "B (LAB)" | "Y (YCbCr)" | "Cb (YCbCr)" | "Cr (YCbCr)") | null; - /** - * Offset - * @description The amount to adjust the channel by - * @default 0 - */ - offset?: number; - /** - * type - * @default img_channel_offset - * @constant - */ - type: "img_channel_offset"; - }; - /** - * Image Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option - */ - ImageCollectionIndexInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Random - * @description Random Index? - * @default true - */ - random?: boolean; - /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) - * @default 0 - */ - index?: number; - /** - * Collection - * @description image collection - * @default null - */ - collection?: components["schemas"]["ImageField"][] | null; - /** - * type - * @default image_collection_index - * @constant - */ - type: "image_collection_index"; - }; - /** - * Image Collection Primitive - * @description A collection of image primitive values - */ - ImageCollectionInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The collection of image values - * @default null - */ - collection?: components["schemas"]["ImageField"][] | null; - /** - * type - * @default image_collection - * @constant - */ - type: "image_collection"; - }; - /** - * Image Collection Primitive linked - * @description A collection of image primitive values - */ - ImageCollectionLinkedInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The collection of image values - * @default null - */ - collection?: components["schemas"]["ImageField"][] | null; - /** - * type - * @default image_collection_linked - * @constant - */ - type: "image_collection_linked"; - /** - * @description The image to load - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - }; - /** - * ImageCollectionOutput - * @description Base class for nodes that output a collection of images - */ - ImageCollectionOutput: { - /** - * Collection - * @description The output images - */ - collection: components["schemas"]["ImageField"][]; - /** - * type - * @default image_collection_output - * @constant - */ - type: "image_collection_output"; - }; - /** - * Image Collection Toggle - * @description Allows boolean selection between two separate Image collection inputs - */ - ImageCollectionToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * Col1 - * @description First Image Collection Input - * @default null - */ - col1?: components["schemas"]["ImageField"][] | null; - /** - * Col2 - * @description Second Image Collection Input - * @default null - */ - col2?: components["schemas"]["ImageField"][] | null; - /** - * type - * @default image_collection_toggle - * @constant - */ - type: "image_collection_toggle"; - }; - /** - * Convert Image Mode - * @description Converts an image to a different mode. - */ - ImageConvertInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to convert - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Mode - * @description The mode to convert to - * @default L - * @enum {string} - */ - mode?: "L" | "RGB" | "RGBA" | "CMYK" | "YCbCr" | "LAB" | "HSV" | "I" | "F"; - /** - * type - * @default img_conv - * @constant - */ - type: "img_conv"; - }; - /** - * Crop Image - * @description Crops an image to a specified box. The box can be outside of the image. - */ - ImageCropInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to crop - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * X - * @description The left x coordinate of the crop rectangle - * @default 0 - */ - x?: number; - /** - * Y - * @description The top y coordinate of the crop rectangle - * @default 0 - */ - y?: number; - /** - * Width - * @description The width of the crop rectangle - * @default 512 - */ - width?: number; - /** - * Height - * @description The height of the crop rectangle - * @default 512 - */ - height?: number; - /** - * type - * @default img_crop - * @constant - */ - type: "img_crop"; - }; - /** - * ImageDTO - * @description Deserialized image record, enriched for the frontend. - */ - ImageDTO: { - /** - * Image Name - * @description The unique name of the image. - */ - image_name: string; - /** - * Image Url - * @description The URL of the image. - */ - image_url: string; - /** - * Thumbnail Url - * @description The URL of the image's thumbnail. - */ - thumbnail_url: string; - /** @description The type of the image. */ - image_origin: components["schemas"]["ResourceOrigin"]; - /** @description The category of the image. */ - image_category: components["schemas"]["ImageCategory"]; - /** - * Width - * @description The width of the image in px. - */ - width: number; - /** - * Height - * @description The height of the image in px. - */ - height: number; - /** - * Created At - * @description The created timestamp of the image. - */ - created_at: string; - /** - * Updated At - * @description The updated timestamp of the image. - */ - updated_at: string; - /** - * Deleted At - * @description The deleted timestamp of the image. - */ - deleted_at?: string | null; - /** - * Is Intermediate - * @description Whether this is an intermediate image. - */ - is_intermediate: boolean; - /** - * Session Id - * @description The session ID that generated this image, if it is a generated image. - */ - session_id?: string | null; - /** - * Node Id - * @description The node ID that generated this image, if it is a generated image. - */ - node_id?: string | null; - /** - * Starred - * @description Whether this image is starred. - */ - starred: boolean; - /** - * Has Workflow - * @description Whether this image has a workflow. - */ - has_workflow: boolean; - /** - * Board Id - * @description The id of the board the image belongs to, if one exists. - */ - board_id?: string | null; - }; - /** - * ImageField - * @description An image primitive field - */ - ImageField: { - /** - * Image Name - * @description The name of the image - */ - image_name: string; - }; - /** - * Image Generator - * @description Generated a collection of images for use in a batched generation - */ - ImageGenerator: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Generator Type - * @description The image generator. - */ - generator: components["schemas"]["ImageGeneratorField"]; - /** - * type - * @default image_generator - * @constant - */ - type: "image_generator"; - }; - /** ImageGeneratorField */ - ImageGeneratorField: Record; - /** - * ImageGeneratorOutput - * @description Base class for nodes that output a collection of boards - */ - ImageGeneratorOutput: { - /** - * Images - * @description The generated images - */ - images: components["schemas"]["ImageField"][]; - /** - * type - * @default image_generator_output - * @constant - */ - type: "image_generator_output"; - }; - /** - * Adjust Image Hue - * @description Adjusts the Hue of an image. - */ - ImageHueAdjustmentInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to adjust - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Hue - * @description The degrees by which to rotate the hue, 0-360 - * @default 0 - */ - hue?: number; - /** - * type - * @default img_hue_adjust - * @constant - */ - type: "img_hue_adjust"; - }; - /** - * Image Index Collect - * @description ImageIndexCollect takes Image and Index then outputs it as an (index,image_name)array converted to json - */ - ImageIndexCollectInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Index - * @description The index - * @default 0 - */ - index?: number; - /** - * @description The image associated with the index - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default image_index_collect - * @constant - */ - type: "image_index_collect"; - }; - /** - * ImageIndexCollectOutput - * @description XImageCollectOutput string containing an array of xItem, Image_name converted to json - */ - ImageIndexCollectOutput: { - /** - * Image Index Collection - * @description The Image Index Collection - */ - image_index_collection: string; - /** - * type - * @default image_index_collect_output - * @constant - */ - type: "image_index_collect_output"; - }; - /** - * Inverse Lerp Image - * @description Inverse linear interpolation of all pixels of an image - */ - ImageInverseLerpInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to lerp - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Min - * @description The minimum input value - * @default 0 - */ - min?: number; - /** - * Max - * @description The maximum input value - * @default 255 - */ - max?: number; - /** - * type - * @default img_ilerp - * @constant - */ - type: "img_ilerp"; - }; - /** - * Image Primitive - * @description An image primitive value - */ - ImageInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to load - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default image - * @constant - */ - type: "image"; - }; - /** - * Lerp Image - * @description Linear interpolation of all pixels of an image - */ - ImageLerpInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to lerp - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Min - * @description The minimum output value - * @default 0 - */ - min?: number; - /** - * Max - * @description The maximum output value - * @default 255 - */ - max?: number; - /** - * type - * @default img_lerp - * @constant - */ - type: "img_lerp"; - }; - /** - * Image Mask to Tensor - * @description Convert a mask image to a tensor. Converts the image to grayscale and uses thresholding at the specified value. - */ - ImageMaskToTensorInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The mask image to convert. - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Cutoff - * @description Cutoff (<) - * @default 128 - */ - cutoff?: number; - /** - * Invert - * @description Whether to invert the mask. - * @default false - */ - invert?: boolean; - /** - * type - * @default image_mask_to_tensor - * @constant - */ - type: "image_mask_to_tensor"; - }; - /** - * Multiply Images - * @description Multiplies two images together using `PIL.ImageChops.multiply()`. - */ - ImageMultiplyInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The first image to multiply - * @default null - */ - image1?: components["schemas"]["ImageField"] | null; - /** - * @description The second image to multiply - * @default null - */ - image2?: components["schemas"]["ImageField"] | null; - /** - * type - * @default img_mul - * @constant - */ - type: "img_mul"; - }; - /** - * Blur NSFW Image - * @description Add blur to NSFW-flagged images - */ - ImageNSFWBlurInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to check - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default img_nsfw - * @constant - */ - type: "img_nsfw"; - }; - /** - * ImageNamesResult - * @description Response containing ordered image names with metadata for optimistic updates. - */ - ImageNamesResult: { - /** - * Image Names - * @description Ordered list of image names - */ - image_names: string[]; - /** - * Starred Count - * @description Number of starred images (when starred_first=True) - */ - starred_count: number; - /** - * Total Count - * @description Total number of images matching the query - */ - total_count: number; - }; - /** - * Add Image Noise - * @description Add noise to an image - */ - ImageNoiseInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to add noise to - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description Optional mask determining where to apply noise (black=noise, white=no noise) - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Seed - * @description Seed for random number generation - * @default 0 - */ - seed?: number; - /** - * Noise Type - * @description The type of noise to add - * @default gaussian - * @enum {string} - */ - noise_type?: "gaussian" | "salt_and_pepper"; - /** - * Amount - * @description The amount of noise to add - * @default 0.1 - */ - amount?: number; - /** - * Noise Color - * @description Whether to add colored noise - * @default true - */ - noise_color?: boolean; - /** - * Size - * @description The size of the noise points - * @default 1 - */ - size?: number; - /** - * type - * @default img_noise - * @constant - */ - type: "img_noise"; - }; - /** - * Offset Image - * @description Offsets an image by a given percentage (or pixel amount). - * - * This works like Offset Latents, but in image space, with the additional capability of taking exact pixel offsets instead of just percentages (toggled with a switch/boolean input). - */ - ImageOffsetInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * As Pixels - * @description Interpret offsets as pixels rather than percentages - * @default false - */ - as_pixels?: boolean; - /** - * @description Image to be offset - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * X Offset - * @description x-offset for the subject - * @default 0.5 - */ - x_offset?: number; - /** - * Y Offset - * @description y-offset for the subject - * @default 0.5 - */ - y_offset?: number; - /** - * type - * @default offset_image - * @constant - */ - type: "offset_image"; - }; - /** - * ImageOutput - * @description Base class for nodes that output a single image - */ - ImageOutput: { - /** @description The output image */ - image: components["schemas"]["ImageField"]; - /** - * Width - * @description The width of the image in pixels - */ - width: number; - /** - * Height - * @description The height of the image in pixels - */ - height: number; - /** - * type - * @default image_output - * @constant - */ - type: "image_output"; - }; - /** ImagePanelCoordinateOutput */ - ImagePanelCoordinateOutput: { - /** - * X Left - * @description The left x-coordinate of the panel. - */ - x_left: number; - /** - * Y Top - * @description The top y-coordinate of the panel. - */ - y_top: number; - /** - * Width - * @description The width of the panel. - */ - width: number; - /** - * Height - * @description The height of the panel. - */ - height: number; - /** - * type - * @default image_panel_coordinate_output - * @constant - */ - type: "image_panel_coordinate_output"; - }; - /** - * Image Panel Layout - * @description Get the coordinates of a single panel in a grid. (If the full image shape cannot be divided evenly into panels, - * then the grid may not cover the entire image.) - */ - ImagePanelLayoutInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Width - * @description The width of the entire grid. - * @default null - */ - width?: number | null; - /** - * Height - * @description The height of the entire grid. - * @default null - */ - height?: number | null; - /** - * Num Cols - * @description The number of columns in the grid. - * @default 1 - */ - num_cols?: number; - /** - * Num Rows - * @description The number of rows in the grid. - * @default 1 - */ - num_rows?: number; - /** - * Panel Col Idx - * @description The column index of the panel to be processed. - * @default 0 - */ - panel_col_idx?: number; - /** - * Panel Row Idx - * @description The row index of the panel to be processed. - * @default 0 - */ - panel_row_idx?: number; - /** - * type - * @default image_panel_layout - * @constant - */ - type: "image_panel_layout"; - }; - /** - * Paste Image - * @description Pastes an image into another image. - */ - ImagePasteInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The base image - * @default null - */ - base_image?: components["schemas"]["ImageField"] | null; - /** - * @description The image to paste - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The mask to use when pasting - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * X - * @description The left x coordinate at which to paste the image - * @default 0 - */ - x?: number; - /** - * Y - * @description The top y coordinate at which to paste the image - * @default 0 - */ - y?: number; - /** - * Crop - * @description Crop to base image dimensions - * @default false - */ - crop?: boolean; - /** - * type - * @default img_paste - * @constant - */ - type: "img_paste"; - }; - /** - * ImageRecordChanges - * @description A set of changes to apply to an image record. - * - * Only limited changes are valid: - * - `image_category`: change the category of an image - * - `session_id`: change the session associated with an image - * - `is_intermediate`: change the image's `is_intermediate` flag - * - `starred`: change whether the image is starred - */ - ImageRecordChanges: { - /** @description The image's new category. */ - image_category?: components["schemas"]["ImageCategory"] | null; - /** - * Session Id - * @description The image's new session ID. - */ - session_id?: string | null; - /** - * Is Intermediate - * @description The image's new `is_intermediate` flag. - */ - is_intermediate?: boolean | null; - /** - * Starred - * @description The image's new `starred` state - */ - starred?: boolean | null; - } & { - [key: string]: unknown; - }; - /** - * Resize Image - * @description Resizes an image to specific dimensions - */ - ImageResizeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to resize - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Width - * @description The width to resize to (px) - * @default 512 - */ - width?: number; - /** - * Height - * @description The height to resize to (px) - * @default 512 - */ - height?: number; - /** - * Resample Mode - * @description The resampling mode - * @default bicubic - * @enum {string} - */ - resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; - /** - * type - * @default img_resize - * @constant - */ - type: "img_resize"; - }; - /** - * Rotate/Flip Image - * @description Rotates an image by a given angle (in degrees clockwise). - * - * Rotate an image in degrees about its center, clockwise (positive entries) or counterclockwise (negative entries). Optionally expand the image boundary to fit the rotated image, or flip it horizontally or vertically. - */ - ImageRotateInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Image to be rotated clockwise - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Degrees - * @description Angle (in degrees clockwise) by which to rotate - * @default 90 - */ - degrees?: number; - /** - * Expand To Fit - * @description If true, extends the image boundary to fit the rotated content - * @default true - */ - expand_to_fit?: boolean; - /** - * Flip Horizontal - * @description If true, flips the image horizontally - * @default false - */ - flip_horizontal?: boolean; - /** - * Flip Vertical - * @description If true, flips the image vertically - * @default false - */ - flip_vertical?: boolean; - /** - * type - * @default rotate_image - * @constant - */ - type: "rotate_image"; - }; - /** - * Image Quantize (Kohonen map) - * @description Use a Kohonen self-organizing map to quantize the pixel values of an image. - * - * It's possible to pass in an existing map, which will be used instead of training a new one. It's also possible to pass in a "swap map", which will be used in place of the standard map's assigned pixel values in quantizing the target image - these values can be correlated either one by one by a linear assignment minimizing the distances\* between each of them, or by swapping their coordinates on the maps themselves, which can be oriented first such that their corner distances\* are minimized achieving a closest-fit while attempting to preserve mappings of adjacent colors. - * - * \*Here, "distances" refers to the euclidean distances between (L, a, b) tuples in Oklab color space. - */ - ImageSOMInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to quantize - * @default null - */ - image_in?: components["schemas"]["ImageField"] | null; - /** - * @description Use an existing SOM instead of training one (skips all training) - * @default null - */ - map_in?: components["schemas"]["ImageField"] | null; - /** - * @description Take another map and swap in its colors after obtaining best-match indices but prior to mapping - * @default null - */ - swap_map?: components["schemas"]["ImageField"] | null; - /** - * Swap Mode - * @description How to employ the swap map - directly, reoriented or rearranged - * @default Minimize distances - * @enum {string} - */ - swap_mode?: "Direct" | "Reorient corners" | "Minimize distances"; - /** - * Map Width - * @description Width (in cells) of the self-organizing map to train - * @default 16 - */ - map_width?: number; - /** - * Map Height - * @description Height (in cells) of the self-organizing map to train - * @default 16 - */ - map_height?: number; - /** - * Steps - * @description Training step count for the self-organizing map - * @default 64 - */ - steps?: number; - /** - * Training Scale - * @description Nearest-neighbor scale image size prior to sampling - size close to sample size is recommended - * @default 0.25 - */ - training_scale?: number; - /** - * Sample Width - * @description Width of assorted pixel sample per step - for performance, keep this number low - * @default 64 - */ - sample_width?: number; - /** - * Sample Height - * @description Height of assorted pixel sample per step - for performance, keep this number low - * @default 64 - */ - sample_height?: number; - /** - * type - * @default image_som - * @constant - */ - type: "image_som"; - }; - /** - * ImageSOMOutput - * @description Outputs an image and the SOM used to quantize that image - */ - ImageSOMOutput: { - /** @description Quantized image */ - image_out: components["schemas"]["ImageField"]; - /** @description The pixels of the self-organizing map */ - map_out: components["schemas"]["ImageField"]; - /** - * Image Width - * @description Width of the quantized image - */ - image_width: number; - /** - * Image Height - * @description Height of the quantized image - */ - image_height: number; - /** - * Map Width - * @description Width of the SOM image - */ - map_width: number; - /** - * Map Height - * @description Height of the SOM image - */ - map_height: number; - /** - * type - * @default image_som_output - * @constant - */ - type: "image_som_output"; - }; - /** - * Scale Image - * @description Scales an image by a factor - */ - ImageScaleInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to scale - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Scale Factor - * @description The factor by which to scale the image - * @default 2 - */ - scale_factor?: number; - /** - * Resample Mode - * @description The resampling mode - * @default bicubic - * @enum {string} - */ - resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; - /** - * type - * @default img_scale - * @constant - */ - type: "img_scale"; - }; - /** - * Image Search to Mask (Clipseg) - * @description Uses the Clipseg model to generate an image mask from an image prompt. - * - * Input a base image and a prompt image to generate a mask representing areas of the base image matched by the prompt image contents. - */ - ImageSearchToMaskClipsegInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image from which to create a mask - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description Image prompt for which to search - * @default null - */ - search_image?: components["schemas"]["ImageField"] | null; - /** - * Invert Output - * @description Off: white on black / On: black on white - * @default true - */ - invert_output?: boolean; - /** - * Smoothing - * @description Radius of blur to apply before thresholding - * @default 4 - */ - smoothing?: number; - /** - * Subject Threshold - * @description Threshold above which is considered the subject - * @default 0.4 - */ - subject_threshold?: number; - /** - * Background Threshold - * @description Threshold below which is considered the background - * @default 0.4 - */ - background_threshold?: number; - /** - * Mask Expand Or Contract - * @description Pixels by which to grow (or shrink) mask after thresholding - * @default 0 - */ - mask_expand_or_contract?: number; - /** - * Mask Blur - * @description Radius of blur to apply after thresholding - * @default 0 - */ - mask_blur?: number; - /** - * type - * @default imgs2mask_clipseg - * @constant - */ - type: "imgs2mask_clipseg"; - }; - /** - * Image to ASCII Art AnyFont - * @description Convert an Image to Ascii Art Image using any font or size - * https://github.com/dernyn/256/tree/master this is a great font to use - */ - ImageToAAInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * @description Image to convert to ASCII art - * @default null - */ - input_image?: components["schemas"]["ImageField"] | null; - /** - * Font Url - * @description URL address of the font file to download - * @default https://github.com/dernyn/256/raw/master/Dernyn's-256(baseline).ttf - */ - font_url?: string | null; - /** - * Local Font Path - * @description Local font file path (overrides font_url) - * @default null - */ - local_font_path?: string | null; - /** - * Local Font - * @description Name of the local font file to use from the font_cache folder - * @default None - * @constant - */ - local_font?: "None"; - /** - * Font Size - * @description Font size for the ASCII art characters - * @default 6 - */ - font_size?: number; - /** - * Character Range - * @description The character range to use - * @default Ascii - * @enum {string} - */ - character_range?: "All" | "Low" | "High" | "Ascii" | "Numbers" | "Letters" | "Lowercase" | "Uppercase" | "Hex" | "Punctuation" | "Printable" | "AH" | "AM" | "AL" | "Blocks" | "Binary" | "Custom"; - /** - * Custom Characters - * @description Custom characters. Used if Custom is selected from character range - * @default ▁▂▃▄▅▆▇█ - */ - custom_characters?: string; - /** - * Comparison Type - * @description Choose the comparison type (Sum of Absolute Differences, Mean Squared Error, Structural Similarity Index, Normalized Average Luminance) - * @default NAL - * @enum {string} - */ - comparison_type?: "SAD" | "MSE" | "SSIM" | "NAL"; - /** - * Mono Comparison - * @description Convert input image to mono for comparison - * @default false - */ - mono_comparison?: boolean; - /** - * Color Mode - * @description Enable color mode (default: grayscale) - * @default false - */ - color_mode?: boolean; - /** - * type - * @default I2AA_AnyFont - * @constant - */ - type: "I2AA_AnyFont"; - }; - /** - * Image to ASCII Art Image - * @description Convert an Image to ASCII Art Image - */ - ImageToDetailedASCIIArtInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * @description Input image to convert to ASCII art - * @default null - */ - input_image?: components["schemas"]["ImageField"] | null; - /** - * Font Spacing - * @description Font spacing for the ASCII art characters - * @default 6 - */ - font_spacing?: number; - /** - * Ascii Set - * @description Choose the desired ASCII character set - * @default Medium Detail - * @enum {string} - */ - ascii_set?: "High Detail" | "Medium Detail" | "Low Detail" | "Numbers" | "Blocks" | "Binary"; - /** - * Color Mode - * @description Enable color mode (default: grayscale) - * @default false - */ - color_mode?: boolean; - /** - * Invert Colors - * @description Invert background color and ASCII character order - * @default true - */ - invert_colors?: boolean; - /** - * Output To File - * @description Output ASCII art to a text file - * @default false - */ - output_to_file?: boolean; - /** - * Gamma - * @description Gamma correction value for the output image - * @default 1 - */ - gamma?: number; - /** - * type - * @default Image_to_ASCII_Art_Image - * @constant - */ - type: "Image_to_ASCII_Art_Image"; - }; - /** - * Image to Image Name - * @description Invocation to extract the image name from an ImageField. - */ - ImageToImageNameInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The ImageField to extract the name from - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default image_to_image_name - * @constant - */ - type: "image_to_image_name"; - }; - /** - * ImageToImageNameOutput - * @description Output class for Image to Image Name Invocation - */ - ImageToImageNameOutput: { - /** - * Image Name - * @description The name of the image - */ - image_name: string; - /** - * type - * @default image_to_image_name_output - * @constant - */ - type: "image_to_image_name_output"; - }; - /** - * Image to Latents - SD1.5, SDXL - * @description Encodes an image into latents. - */ - ImageToLatentsInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to encode - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description VAE - * @default null - */ - vae?: components["schemas"]["VAEField"] | null; - /** - * Tiled - * @description Processing using overlapping tiles (reduce memory consumption) - * @default false - */ - tiled?: boolean; - /** - * Tile Size - * @description The tile size for VAE tiling in pixels (image space). If set to 0, the default tile size for the model will be used. Larger tile sizes generally produce better results at the cost of higher memory usage. - * @default 0 - */ - tile_size?: number; - /** - * Fp32 - * @description Whether or not to use full float32 precision - * @default false - */ - fp32?: boolean; - /** - * Color Compensation - * @description Apply VAE scaling compensation when encoding images (reduces color drift). - * @default None - * @enum {string} - */ - color_compensation?: "None" | "SDXL"; - /** - * type - * @default i2l - * @constant - */ - type: "i2l"; - }; - /** - * Image to Unicode Art - * @description Convert an Image to Unicode Art using Extended Characters - */ - ImageToUnicodeArtInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * @description Input image to convert to Unicode art - * @default null - */ - input_image?: components["schemas"]["ImageField"] | null; - /** - * Font Size - * @description Font size for the Unicode art characters - * @default 8 - */ - font_size?: number; - /** - * Gamma - * @description Gamma correction value for the output image - * @default 1 - */ - gamma?: number; - /** - * Unicode Set - * @description Use shaded Unicode characters for artwork - * @default Shaded - * @enum {string} - */ - unicode_set?: "Shaded" | "Extended Shading" | "Intermediate Detail" | "Checkerboard Patterns" | "Vertical Lines" | "Horizontal Lines" | "Diagonal Lines" | "Arrows" | "Circles" | "Blocks" | "Triangles" | "Math Symbols" | "Stars"; - /** - * Color Mode - * @description Enable color mode (default: grayscale) - * @default true - */ - color_mode?: boolean; - /** - * Invert Colors - * @description Invert background color and ASCII character order - * @default true - */ - invert_colors?: boolean; - /** - * type - * @default Image_to_Unicode_Art - * @constant - */ - type: "Image_to_Unicode_Art"; - }; - /** - * Image To XYImage Collection - * @description Cuts an image up into columns and rows and outputs XYImage Collection - */ - ImageToXYImageCollectionInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The input image - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Columns - * @description The number of columns - * @default 2 - */ - columns?: number; - /** - * Rows - * @description The number of rows - * @default 2 - */ - rows?: number; - /** - * type - * @default image_to_xy_image_collection - * @constant - */ - type: "image_to_xy_image_collection"; - }; - /** - * Image To XYImage Tiles - * @description Cuts an image up into overlapping tiles and outputs as an XYImage Collection (x,y is the final position of the tile) - */ - ImageToXYImageTilesInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Tiles - * @description The list of tiles - * @default [] - */ - tiles?: string[]; - /** - * type - * @default image_to_xy_image_tiles - * @constant - */ - type: "image_to_xy_image_tiles"; - }; - /** - * ImageToXYImageTilesOutput - * @description Image To XYImage Tiles Output - */ - ImageToXYImageTilesOutput: { - /** - * Xyimages - * @description The XYImage Collection - */ - xyImages: string[]; - /** - * type - * @default image_to_xy_image_output - * @constant - */ - type: "image_to_xy_image_output"; - }; - /** - * Image Toggle - * @description Allows boolean selection between two separate image inputs - */ - ImageToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * @description First Image Input - * @default null - */ - img1?: components["schemas"]["ImageField"] | null; - /** - * @description Second Image Input - * @default null - */ - img2?: components["schemas"]["ImageField"] | null; - /** - * type - * @default image_toggle - * @constant - */ - type: "image_toggle"; - }; - /** ImageUploadEntry */ - ImageUploadEntry: { - /** @description The image DTO */ - image_dto: components["schemas"]["ImageDTO"]; - /** - * Presigned Url - * @description The URL to get the presigned URL for the image upload - */ - presigned_url: string; - }; - /** - * ImageUrlsDTO - * @description The URLs for an image and its thumbnail. - */ - ImageUrlsDTO: { - /** - * Image Name - * @description The unique name of the image. - */ - image_name: string; - /** - * Image Url - * @description The URL of the image. - */ - image_url: string; - /** - * Thumbnail Url - * @description The URL of the image's thumbnail. - */ - thumbnail_url: string; - }; - /** - * Add Invisible Watermark - * @description Add an invisible watermark to an image - */ - ImageWatermarkInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to check - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Text - * @description Watermark text - * @default InvokeAI - */ - text?: string; - /** - * type - * @default img_watermark - * @constant - */ - type: "img_watermark"; - }; - /** ImagesDownloaded */ - ImagesDownloaded: { - /** - * Response - * @description The message to display to the user when images begin downloading - */ - response?: string | null; - /** - * Bulk Download Item Name - * @description The name of the bulk download item for which events will be emitted - */ - bulk_download_item_name?: string | null; - }; - /** - * Images Index To Video Output - * @description Load a collection of ImageIndex types (json of (idex,image_name)array) and create a video of them - */ - ImagesIndexToVideoInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Image Index Collection - * @description The Image Index Collection - * @default null - */ - image_index_collection?: string[] | null; - /** - * Video Out Path - * @description Path and filename of output mp4 - * @default - */ - video_out_path?: string; - /** - * Fps - * @description FPS - * @default 30 - */ - fps?: number; - /** - * Codec - * @description Video codec FourCC (e.g., mp4v, avc1, h264, x264, hev1, vp09, av01) - * @default x264 - */ - codec?: string; - /** - * type - * @default images_index_to_video - * @constant - */ - type: "images_index_to_video"; - }; - /** - * ImagesIndexToVideoOutput - * @description ImagesIndexToVideoOutput returns nothing - */ - ImagesIndexToVideoOutput: { - /** - * type - * @default ImagesIndexToVideoOutput - * @constant - */ - type: "ImagesIndexToVideoOutput"; - }; - /** - * Images To Grids - * @description Takes a collection of images and outputs a collection of generated grid images - */ - ImagesToGridsInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Images - * @description The image collection to turn into grids - * @default [] - */ - images?: components["schemas"]["ImageField"][]; - /** - * Columns - * @description The number of columns in each grid - * @default 1 - */ - columns?: number; - /** - * Rows - * @description The number of rows to have in each grid - * @default 1 - */ - rows?: number; - /** - * Space - * @description The space to be added between images - * @default 1 - */ - space?: number; - /** - * Scale Factor - * @description The factor by which to scale the images - * @default 1 - */ - scale_factor?: number; - /** - * Resample Mode - * @description The resampling mode - * @default bicubic - * @enum {string} - */ - resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; - /** - * @description The color to use as the background - * @default { - * "r": 0, - * "g": 0, - * "b": 0, - * "a": 255 - * } - */ - background_color?: components["schemas"]["ColorField"]; - /** - * type - * @default images_to_grids - * @constant - */ - type: "images_to_grids"; - }; - /** - * Solid Color Infill - * @description Infills transparent areas of an image with a solid color - */ - InfillColorInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The color to use to infill - * @default { - * "r": 127, - * "g": 127, - * "b": 127, - * "a": 255 - * } - */ - color?: components["schemas"]["ColorField"]; - /** - * type - * @default infill_rgba - * @constant - */ - type: "infill_rgba"; - }; - /** - * PatchMatch Infill - * @description Infills transparent areas of an image using the PatchMatch algorithm - */ - InfillPatchMatchInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Downscale - * @description Run patchmatch on downscaled image to speedup infill - * @default 2 - */ - downscale?: number; - /** - * Resample Mode - * @description The resampling mode - * @default bicubic - * @enum {string} - */ - resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; - /** - * type - * @default infill_patchmatch - * @constant - */ - type: "infill_patchmatch"; - }; - /** - * Tile Infill - * @description Infills transparent areas of an image with tiles of the image - */ - InfillTileInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Tile Size - * @description The tile size (px) - * @default 32 - */ - tile_size?: number; - /** - * Seed - * @description The seed to use for tile generation (omit for random) - * @default 0 - */ - seed?: number; - /** - * type - * @default infill_tile - * @constant - */ - type: "infill_tile"; - }; - /** - * UNet Info Grabber - * @description Outputs different info from a UNet - */ - InfoGrabberUNetInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * UNet - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; - /** - * Info Type - * @description The kind of info to retrieve - * @default null - */ - info_type?: ("name" | "path") | null; - /** - * type - * @default info_grabber_unet_invocation - * @constant - */ - type: "info_grabber_unet_invocation"; - }; - /** - * Input - * @description The type of input a field accepts. - * - `Input.Direct`: The field must have its value provided directly, when the invocation and field are instantiated. - * - `Input.Connection`: The field must have its value provided by a connection. - * - `Input.Any`: The field may have its value provided either directly or by a connection. - * @enum {string} - */ - Input: "connection" | "direct" | "any"; - /** - * InputFieldJSONSchemaExtra - * @description Extra attributes to be added to input fields and their OpenAPI schema. Used during graph execution, - * and by the workflow editor during schema parsing and UI rendering. - */ - InputFieldJSONSchemaExtra: { - input: components["schemas"]["Input"]; - field_kind: components["schemas"]["FieldKind"]; - /** - * Orig Required - * @default true - */ - orig_required: boolean; - /** - * Default - * @default null - */ - default: unknown | null; - /** - * Orig Default - * @default null - */ - orig_default: unknown | null; - /** - * Ui Hidden - * @default false - */ - ui_hidden: boolean; - /** @default null */ - ui_type: components["schemas"]["UIType"] | null; - /** @default null */ - ui_component: components["schemas"]["UIComponent"] | null; - /** - * Ui Order - * @default null - */ - ui_order: number | null; - /** - * Ui Choice Labels - * @default null - */ - ui_choice_labels: { - [key: string]: string; - } | null; - /** - * Ui Model Base - * @default null - */ - ui_model_base: components["schemas"]["BaseModelType"][] | null; - /** - * Ui Model Type - * @default null - */ - ui_model_type: components["schemas"]["ModelType"][] | null; - /** - * Ui Model Variant - * @default null - */ - ui_model_variant: (components["schemas"]["ClipVariantType"] | components["schemas"]["ModelVariantType"])[] | null; - /** - * Ui Model Format - * @default null - */ - ui_model_format: components["schemas"]["ModelFormat"][] | null; - }; - /** - * InstallStatus - * @description State of an install job running in the background. - * @enum {string} - */ - InstallStatus: "waiting" | "downloading" | "downloads_done" | "running" | "paused" | "completed" | "error" | "cancelled"; - /** - * Integer Collection Toggle - * @description Allows boolean selection between two separate integer collection inputs - */ - IntCollectionToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * Col1 - * @description First Int Collection Input - * @default null - */ - col1?: number[] | null; - /** - * Col2 - * @description Second Int Collection Input - * @default null - */ - col2?: number[] | null; - /** - * type - * @default int_collection_toggle - * @constant - */ - type: "int_collection_toggle"; - }; - /** - * Integer Toggle - * @description Allows boolean selection between two separate integer inputs - */ - IntToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * Int1 - * @description First Integer Input - * @default 0 - */ - int1?: number; - /** - * Int2 - * @description Second Integer Input - * @default 0 - */ - int2?: number; - /** - * type - * @default int_toggle - * @constant - */ - type: "int_toggle"; - }; - /** - * Integer Batch - * @description Create a batched generation, where the workflow is executed once for each integer in the batch. - */ - IntegerBatchInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Batch Group - * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. - * @default None - * @enum {string} - */ - batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; - /** - * Integers - * @description The integers to batch over - * @default null - */ - integers?: number[] | null; - /** - * type - * @default integer_batch - * @constant - */ - type: "integer_batch"; - }; - /** - * Integer Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option - */ - IntegerCollectionIndexInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Random - * @description Random Index? - * @default true - */ - random?: boolean; - /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) - * @default 0 - */ - index?: number; - /** - * Collection - * @description integer collection - * @default null - */ - collection?: number[] | null; - /** - * type - * @default integer_collection_index - * @constant - */ - type: "integer_collection_index"; - }; - /** - * Integer Collection Primitive - * @description A collection of integer primitive values - */ - IntegerCollectionInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The collection of integer values - * @default [] - */ - collection?: number[]; - /** - * type - * @default integer_collection - * @constant - */ - type: "integer_collection"; - }; - /** - * Integer Collection Primitive Linked - * @description A collection of integer primitive values - */ - IntegerCollectionLinkedInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The collection of integer values - * @default [] - */ - collection?: number[]; - /** - * type - * @default integer_collection_linked - * @constant - */ - type: "integer_collection_linked"; - /** - * Value - * @description The integer value - * @default 0 - */ - value?: number; - }; - /** - * IntegerCollectionOutput - * @description Base class for nodes that output a collection of integers - */ - IntegerCollectionOutput: { - /** - * Collection - * @description The int collection - */ - collection: number[]; - /** - * type - * @default integer_collection_output - * @constant - */ - type: "integer_collection_output"; - }; - /** - * Integer Generator - * @description Generated a range of integers for use in a batched generation - */ - IntegerGenerator: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Generator Type - * @description The integer generator. - */ - generator: components["schemas"]["IntegerGeneratorField"]; - /** - * type - * @default integer_generator - * @constant - */ - type: "integer_generator"; - }; - /** IntegerGeneratorField */ - IntegerGeneratorField: Record; - /** IntegerGeneratorOutput */ - IntegerGeneratorOutput: { - /** - * Integers - * @description The generated integers - */ - integers: number[]; - /** - * type - * @default integer_generator_output - * @constant - */ - type: "integer_generator_output"; - }; - /** - * Integer Primitive - * @description An integer primitive value - */ - IntegerInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Value - * @description The integer value - * @default 0 - */ - value?: number; - /** - * type - * @default integer - * @constant - */ - type: "integer"; - }; - /** - * Integer Math - * @description Performs integer math. - */ - IntegerMathInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Operation - * @description The operation to perform - * @default ADD - * @enum {string} - */ - operation?: "ADD" | "SUB" | "MUL" | "DIV" | "EXP" | "MOD" | "ABS" | "MIN" | "MAX"; - /** - * A - * @description The first number - * @default 1 - */ - a?: number; - /** - * B - * @description The second number - * @default 1 - */ - b?: number; - /** - * type - * @default integer_math - * @constant - */ - type: "integer_math"; - }; - /** - * IntegerOutput - * @description Base class for nodes that output a single integer - */ - IntegerOutput: { - /** - * Value - * @description The output integer - */ - value: number; - /** - * type - * @default integer_output - * @constant - */ - type: "integer_output"; - }; - /** - * Ints To Strings - * @description Converts an integer or collection of integers to a collection of strings - */ - IntsToStringsInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Ints - * @description int or collection of ints - * @default [] - */ - ints?: number | number[]; - /** - * type - * @default ints_to_strings - * @constant - */ - type: "ints_to_strings"; - }; - /** - * Invert Tensor Mask - * @description Inverts a tensor mask. - */ - InvertTensorMaskInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The tensor mask to convert. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default invert_tensor_mask - * @constant - */ - type: "invert_tensor_mask"; - }; - /** InvocationCacheStatus */ - InvocationCacheStatus: { - /** - * Size - * @description The current size of the invocation cache - */ - size: number; - /** - * Hits - * @description The number of cache hits - */ - hits: number; - /** - * Misses - * @description The number of cache misses - */ - misses: number; - /** - * Enabled - * @description Whether the invocation cache is enabled - */ - enabled: boolean; - /** - * Max Size - * @description The maximum size of the invocation cache - */ - max_size: number; - }; - /** - * InvocationCompleteEvent - * @description Event model for invocation_complete - */ - InvocationCompleteEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; - /** - * Queue Id - * @description The ID of the queue - */ - queue_id: string; - /** - * Item Id - * @description The ID of the queue item - */ - item_id: number; - /** - * Batch Id - * @description The ID of the queue batch - */ - batch_id: string; - /** - * Origin - * @description The origin of the queue item - * @default null - */ - origin: string | null; - /** - * Destination - * @description The destination of the queue item - * @default null - */ - destination: string | null; - /** - * User Id - * @description The ID of the user who created the queue item - * @default system - */ - user_id: string; - /** - * Session Id - * @description The ID of the session (aka graph execution state) - */ - session_id: string; - /** - * Invocation - * @description The ID of the invocation - */ - invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; - /** - * Invocation Source Id - * @description The ID of the prepared invocation's source node - */ - invocation_source_id: string; - /** - * Result - * @description The result of the invocation - */ - result: components["schemas"]["BooleanCollectionOutput"] | components["schemas"]["BooleanOutput"] | components["schemas"]["BoundingBoxCollectionOutput"] | components["schemas"]["BoundingBoxOutput"] | components["schemas"]["CLIPOutput"] | components["schemas"]["CLIPSkipInvocationOutput"] | components["schemas"]["CMYKSeparationOutput"] | components["schemas"]["CMYKSplitOutput"] | components["schemas"]["CalculateImageTilesOutput"] | components["schemas"]["ClipsegMaskHierarchyOutput"] | components["schemas"]["CogView4ConditioningOutput"] | components["schemas"]["CogView4ModelLoaderOutput"] | components["schemas"]["CollectInvocationOutput"] | components["schemas"]["CollectionCountOutput"] | components["schemas"]["CollectionIndexOutput"] | components["schemas"]["CollectionJoinOutput"] | components["schemas"]["CollectionReverseOutput"] | components["schemas"]["CollectionSliceOutput"] | components["schemas"]["CollectionSortOutput"] | components["schemas"]["CollectionUniqueOutput"] | components["schemas"]["ColorCollectionOutput"] | components["schemas"]["ColorOutput"] | components["schemas"]["ConditioningCollectionOutput"] | components["schemas"]["ConditioningOutput"] | components["schemas"]["ControlListOutput"] | components["schemas"]["ControlOutput"] | components["schemas"]["DenoiseMaskOutput"] | components["schemas"]["EscapedOutput"] | components["schemas"]["EvolutionaryPromptListOutput"] | components["schemas"]["FaceMaskOutput"] | components["schemas"]["FaceOffOutput"] | components["schemas"]["FloatCollectionOutput"] | components["schemas"]["FloatGeneratorOutput"] | components["schemas"]["FloatOutput"] | components["schemas"]["Flux2KleinLoRALoaderOutput"] | components["schemas"]["Flux2KleinModelLoaderOutput"] | components["schemas"]["FluxConditioningBlendOutput"] | components["schemas"]["FluxConditioningCollectionOutput"] | components["schemas"]["FluxConditioningDeltaAndAugmentedOutput"] | components["schemas"]["FluxConditioningListOutput"] | components["schemas"]["FluxConditioningOutput"] | components["schemas"]["FluxConditioningStoreOutput"] | components["schemas"]["FluxControlLoRALoaderOutput"] | components["schemas"]["FluxControlNetCollectionOutput"] | components["schemas"]["FluxControlNetListOutput"] | components["schemas"]["FluxControlNetOutput"] | components["schemas"]["FluxFillOutput"] | components["schemas"]["FluxIdealSizeOutput"] | components["schemas"]["FluxKontextOutput"] | components["schemas"]["FluxLoRALoaderOutput"] | components["schemas"]["FluxModelLoaderOutput"] | components["schemas"]["FluxReduxCollectionOutput"] | components["schemas"]["FluxReduxListOutput"] | components["schemas"]["FluxReduxOutput"] | components["schemas"]["FluxTextEncoderListOutput"] | components["schemas"]["FluxWeightedPromptOutput"] | components["schemas"]["GradientMaskOutput"] | components["schemas"]["HalvedPromptOutput"] | components["schemas"]["IPAdapterListOutput"] | components["schemas"]["IPAdapterOutput"] | components["schemas"]["IdealSizeOutput"] | components["schemas"]["ImageCollectionOutput"] | components["schemas"]["ImageGeneratorOutput"] | components["schemas"]["ImageIndexCollectOutput"] | components["schemas"]["ImageOutput"] | components["schemas"]["ImagePanelCoordinateOutput"] | components["schemas"]["ImageSOMOutput"] | components["schemas"]["ImageToImageNameOutput"] | components["schemas"]["ImageToXYImageTilesOutput"] | components["schemas"]["ImagesIndexToVideoOutput"] | components["schemas"]["IntegerCollectionOutput"] | components["schemas"]["IntegerGeneratorOutput"] | components["schemas"]["IntegerOutput"] | components["schemas"]["IterateInvocationOutput"] | components["schemas"]["JsonListStringsOutput"] | components["schemas"]["LatentsCollectionOutput"] | components["schemas"]["LatentsMetaOutput"] | components["schemas"]["LatentsOutput"] | components["schemas"]["LoRACollectionFromPathOutput"] | components["schemas"]["LoRACollectionOutput"] | components["schemas"]["LoRACollectionToggleOutput"] | components["schemas"]["LoRALoaderOutput"] | components["schemas"]["LoRANameGrabberOutput"] | components["schemas"]["LoRASelectorOutput"] | components["schemas"]["LoRAToggleOutput"] | components["schemas"]["LoadAllTextFilesInFolderOutput"] | components["schemas"]["LoadTextFileToStringOutput"] | components["schemas"]["LookupLoRACollectionTriggersOutput"] | components["schemas"]["LookupLoRATriggersOutput"] | components["schemas"]["LookupTableOutput"] | components["schemas"]["MDControlListOutput"] | components["schemas"]["MDIPAdapterListOutput"] | components["schemas"]["MDT2IAdapterListOutput"] | components["schemas"]["MaskOutput"] | components["schemas"]["MathEvalOutput"] | components["schemas"]["MergeLoRACollectionsOutput"] | components["schemas"]["MergeStringCollectionsOutput"] | components["schemas"]["MetadataItemOutput"] | components["schemas"]["MetadataOutput"] | components["schemas"]["MetadataToLorasCollectionOutput"] | components["schemas"]["MetadataToModelOutput"] | components["schemas"]["MetadataToSDXLModelOutput"] | components["schemas"]["ModelIdentifierOutput"] | components["schemas"]["ModelLoaderOutput"] | components["schemas"]["ModelNameGrabberOutput"] | components["schemas"]["ModelToggleOutput"] | components["schemas"]["NightmareOutput"] | components["schemas"]["NoiseOutput"] | components["schemas"]["PBRMapsOutput"] | components["schemas"]["PTFieldsCollectOutput"] | components["schemas"]["PTFieldsExpandOutput"] | components["schemas"]["PairTileImageOutput"] | components["schemas"]["PaletteOutput"] | components["schemas"]["PrintStringToConsoleOutput"] | components["schemas"]["PromptStrengthOutput"] | components["schemas"]["PromptTemplateOutput"] | components["schemas"]["PromptsToFileInvocationOutput"] | components["schemas"]["PrunedPromptOutput"] | components["schemas"]["Qwen3PromptProOutput"] | components["schemas"]["RGBSplitOutput"] | components["schemas"]["RandomImageSizeOutput"] | components["schemas"]["RandomLoRAMixerOutput"] | components["schemas"]["ReapplyLoRAWeightOutput"] | components["schemas"]["RetrieveFluxConditioningMultiOutput"] | components["schemas"]["SD3ConditioningOutput"] | components["schemas"]["SDXLLoRALoaderOutput"] | components["schemas"]["SDXLModelLoaderOutput"] | components["schemas"]["SDXLRefinerModelLoaderOutput"] | components["schemas"]["SchedulerOutput"] | components["schemas"]["Sd3ModelLoaderOutput"] | components["schemas"]["SeamlessModeOutput"] | components["schemas"]["ShadowsHighlightsMidtonesMasksOutput"] | components["schemas"]["String2Output"] | components["schemas"]["StringCollectionJoinerOutput"] | components["schemas"]["StringCollectionOutput"] | components["schemas"]["StringGeneratorOutput"] | components["schemas"]["StringOutput"] | components["schemas"]["StringPosNegOutput"] | components["schemas"]["StringToLoraOutput"] | components["schemas"]["StringToMainModelOutput"] | components["schemas"]["StringToModelOutput"] | components["schemas"]["StringToSDXLModelOutput"] | components["schemas"]["T2IAdapterListOutput"] | components["schemas"]["T2IAdapterOutput"] | components["schemas"]["ThresholdingOutput"] | components["schemas"]["TileSizeOutput"] | components["schemas"]["TileToPropertiesOutput"] | components["schemas"]["TilesOutput"] | components["schemas"]["TraceryOutput"] | components["schemas"]["UNetOutput"] | components["schemas"]["VAEOutput"] | components["schemas"]["WeightedStringOutput"] | components["schemas"]["XYExpandOutput"] | components["schemas"]["XYImageExpandOutput"] | components["schemas"]["XYProductOutput"] | components["schemas"]["ZImageConditioningOutput"] | components["schemas"]["ZImageControlOutput"] | components["schemas"]["ZImageLoRALoaderOutput"] | components["schemas"]["ZImageModelLoaderOutput"]; - }; - /** - * InvocationErrorEvent - * @description Event model for invocation_error - */ - InvocationErrorEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; - /** - * Queue Id - * @description The ID of the queue - */ - queue_id: string; - /** - * Item Id - * @description The ID of the queue item - */ - item_id: number; - /** - * Batch Id - * @description The ID of the queue batch - */ - batch_id: string; - /** - * Origin - * @description The origin of the queue item - * @default null - */ - origin: string | null; - /** - * Destination - * @description The destination of the queue item - * @default null - */ - destination: string | null; - /** - * User Id - * @description The ID of the user who created the queue item - * @default system - */ - user_id: string; - /** - * Session Id - * @description The ID of the session (aka graph execution state) - */ - session_id: string; - /** - * Invocation - * @description The ID of the invocation - */ - invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; - /** - * Invocation Source Id - * @description The ID of the prepared invocation's source node - */ - invocation_source_id: string; - /** - * Error Type - * @description The error type - */ - error_type: string; - /** - * Error Message - * @description The error message - */ - error_message: string; - /** - * Error Traceback - * @description The error traceback - */ - error_traceback: string; - }; - InvocationOutputMap: { - Advanced_Text_Font_to_Image: components["schemas"]["ImageOutput"]; - I2AA_AnyFont: components["schemas"]["ImageOutput"]; - Image_to_ASCII_Art_Image: components["schemas"]["ImageOutput"]; - Image_to_Unicode_Art: components["schemas"]["ImageOutput"]; - Retrieve_Board_Images: components["schemas"]["ImageCollectionOutput"]; - Text_Font_to_Image: components["schemas"]["ImageOutput"]; - add: components["schemas"]["IntegerOutput"]; - adv_autostereogram: components["schemas"]["ImageOutput"]; - alpha_mask_to_tensor: components["schemas"]["MaskOutput"]; - anamorphic_streaks: components["schemas"]["ImageOutput"]; - apply_mask_to_image: components["schemas"]["ImageOutput"]; - apply_tensor_mask_to_image: components["schemas"]["ImageOutput"]; - autostereogram: components["schemas"]["ImageOutput"]; - average_images: components["schemas"]["ImageOutput"]; - blank_image: components["schemas"]["ImageOutput"]; - bool_collection_index: components["schemas"]["BooleanOutput"]; - bool_collection_toggle: components["schemas"]["BooleanCollectionOutput"]; - bool_toggle: components["schemas"]["BooleanOutput"]; - boolean: components["schemas"]["BooleanOutput"]; - boolean_collection: components["schemas"]["BooleanCollectionOutput"]; - boolean_collection_linked: components["schemas"]["BooleanCollectionOutput"]; - bounding_box: components["schemas"]["BoundingBoxOutput"]; - calculate_image_tiles: components["schemas"]["CalculateImageTilesOutput"]; - calculate_image_tiles_even_split: components["schemas"]["CalculateImageTilesOutput"]; - calculate_image_tiles_min_overlap: components["schemas"]["CalculateImageTilesOutput"]; - canny_edge_detection: components["schemas"]["ImageOutput"]; - canvas_paste_back: components["schemas"]["ImageOutput"]; - canvas_v2_mask_and_crop: components["schemas"]["ImageOutput"]; - chroma_noise: components["schemas"]["ImageOutput"]; - chromatic_aberration: components["schemas"]["ImageOutput"]; - clip_skip: components["schemas"]["CLIPSkipInvocationOutput"]; - clipseg_mask_hierarchy: components["schemas"]["ClipsegMaskHierarchyOutput"]; - cmyk_halftone: components["schemas"]["ImageOutput"]; - cmyk_merge: components["schemas"]["ImageOutput"]; - cmyk_separation: components["schemas"]["CMYKSeparationOutput"]; - cmyk_split: components["schemas"]["CMYKSplitOutput"]; - cogview4_denoise: components["schemas"]["LatentsOutput"]; - cogview4_i2l: components["schemas"]["LatentsOutput"]; - cogview4_l2i: components["schemas"]["ImageOutput"]; - cogview4_model_loader: components["schemas"]["CogView4ModelLoaderOutput"]; - cogview4_text_encoder: components["schemas"]["CogView4ConditioningOutput"]; - collect: components["schemas"]["CollectInvocationOutput"]; - collection_count: components["schemas"]["CollectionCountOutput"]; - collection_index: components["schemas"]["CollectionIndexOutput"]; - collection_join: components["schemas"]["CollectionJoinOutput"]; - collection_reverse: components["schemas"]["CollectionReverseOutput"]; - collection_slice: components["schemas"]["CollectionSliceOutput"]; - collection_sort: components["schemas"]["CollectionSortOutput"]; - collection_unique: components["schemas"]["CollectionUniqueOutput"]; - color: components["schemas"]["ColorOutput"]; - color_cast_correction: components["schemas"]["ImageOutput"]; - color_correct: components["schemas"]["ImageOutput"]; - color_map: components["schemas"]["ImageOutput"]; - compare_floats_invocation: components["schemas"]["BooleanOutput"]; - compare_ints_invocation: components["schemas"]["BooleanOutput"]; - compare_strings_invocation: components["schemas"]["BooleanOutput"]; - compel: components["schemas"]["ConditioningOutput"]; - conditioning: components["schemas"]["ConditioningOutput"]; - conditioning_collection: components["schemas"]["ConditioningCollectionOutput"]; - conditioning_collection_linked: components["schemas"]["ConditioningCollectionOutput"]; - conditioning_collection_toggle: components["schemas"]["ConditioningCollectionOutput"]; - conditioning_toggle: components["schemas"]["ConditioningOutput"]; - content_shuffle: components["schemas"]["ImageOutput"]; - controlnet: components["schemas"]["ControlOutput"]; - "controlnet-linked": components["schemas"]["ControlListOutput"]; - core_metadata: components["schemas"]["MetadataOutput"]; - create_denoise_mask: components["schemas"]["DenoiseMaskOutput"]; - create_gradient_mask: components["schemas"]["GradientMaskOutput"]; - crop_image_to_bounding_box: components["schemas"]["ImageOutput"]; - crop_latents: components["schemas"]["LatentsOutput"]; - crossover_prompt: components["schemas"]["HalvedPromptOutput"]; - csv_to_index_string: components["schemas"]["StringOutput"]; - csv_to_strings: components["schemas"]["StringCollectionOutput"]; - cv_extrude_depth: components["schemas"]["ImageOutput"]; - cv_inpaint: components["schemas"]["ImageOutput"]; - decode_watermark: components["schemas"]["StringOutput"]; - default_xy_tile_generator: components["schemas"]["TilesOutput"]; - denoise_latents: components["schemas"]["LatentsOutput"]; - denoise_latents_meta: components["schemas"]["LatentsMetaOutput"]; - depth_anything_depth_estimation: components["schemas"]["ImageOutput"]; - div: components["schemas"]["IntegerOutput"]; - dw_openpose_detection: components["schemas"]["ImageOutput"]; - dynamic_prompt: components["schemas"]["StringCollectionOutput"]; - enhance_detail: components["schemas"]["ImageOutput"]; - esrgan: components["schemas"]["ImageOutput"]; - even_split_xy_tile_generator: components["schemas"]["TilesOutput"]; - expand_mask_with_fade: components["schemas"]["ImageOutput"]; - extract_image_collection_metadata_boolean: components["schemas"]["BooleanCollectionOutput"]; - extract_image_collection_metadata_float: components["schemas"]["FloatCollectionOutput"]; - extract_image_collection_metadata_integer: components["schemas"]["IntegerCollectionOutput"]; - extract_image_collection_metadata_string: components["schemas"]["StringCollectionOutput"]; - face_identifier: components["schemas"]["ImageOutput"]; - face_mask_detection: components["schemas"]["FaceMaskOutput"]; - face_off: components["schemas"]["FaceOffOutput"]; - film_grain: components["schemas"]["ImageOutput"]; - flatten_histogram_mono: components["schemas"]["ImageOutput"]; - float: components["schemas"]["FloatOutput"]; - float_batch: components["schemas"]["FloatOutput"]; - float_collection: components["schemas"]["FloatCollectionOutput"]; - float_collection_index: components["schemas"]["FloatOutput"]; - float_collection_linked: components["schemas"]["FloatCollectionOutput"]; - float_collection_toggle: components["schemas"]["FloatCollectionOutput"]; - float_generator: components["schemas"]["FloatGeneratorOutput"]; - float_math: components["schemas"]["FloatOutput"]; - float_range: components["schemas"]["FloatCollectionOutput"]; - float_to_int: components["schemas"]["IntegerOutput"]; - float_toggle: components["schemas"]["FloatOutput"]; - floats_to_strings: components["schemas"]["StringCollectionOutput"]; - flux2_denoise: components["schemas"]["LatentsOutput"]; - flux2_klein_lora_collection_loader: components["schemas"]["Flux2KleinLoRALoaderOutput"]; - flux2_klein_lora_loader: components["schemas"]["Flux2KleinLoRALoaderOutput"]; - flux2_klein_model_loader: components["schemas"]["Flux2KleinModelLoaderOutput"]; - flux2_klein_text_encoder: components["schemas"]["FluxConditioningOutput"]; - flux2_vae_decode: components["schemas"]["ImageOutput"]; - flux2_vae_encode: components["schemas"]["LatentsOutput"]; - flux_conditioning_blend: components["schemas"]["FluxConditioningBlendOutput"]; - flux_conditioning_collection: components["schemas"]["FluxConditioningCollectionOutput"]; - flux_conditioning_collection_join: components["schemas"]["FluxConditioningCollectionOutput"]; - flux_conditioning_collection_toggle: components["schemas"]["FluxConditioningCollectionOutput"]; - flux_conditioning_concatenate: components["schemas"]["FluxConditioningOutput"]; - flux_conditioning_delta_augmentation: components["schemas"]["FluxConditioningDeltaAndAugmentedOutput"]; - flux_conditioning_index: components["schemas"]["FluxConditioningOutput"]; - flux_conditioning_list: components["schemas"]["FluxConditioningListOutput"]; - flux_conditioning_math: components["schemas"]["FluxConditioningOutput"]; - flux_conditioning_toggle: components["schemas"]["FluxConditioningOutput"]; - flux_control_lora_loader: components["schemas"]["FluxControlLoRALoaderOutput"]; - flux_controlnet: components["schemas"]["FluxControlNetOutput"]; - flux_controlnet_collection: components["schemas"]["FluxControlNetCollectionOutput"]; - flux_controlnet_collection_join: components["schemas"]["FluxControlNetCollectionOutput"]; - flux_controlnet_index: components["schemas"]["FluxControlNetOutput"]; - flux_controlnet_linked: components["schemas"]["FluxControlNetListOutput"]; - flux_denoise: components["schemas"]["LatentsOutput"]; - flux_denoise_meta: components["schemas"]["LatentsMetaOutput"]; - flux_fill: components["schemas"]["FluxFillOutput"]; - flux_ideal_size: components["schemas"]["FluxIdealSizeOutput"]; - flux_ip_adapter: components["schemas"]["IPAdapterOutput"]; - flux_kontext: components["schemas"]["FluxKontextOutput"]; - flux_kontext_ideal_size: components["schemas"]["FluxIdealSizeOutput"]; - flux_kontext_image_prep: components["schemas"]["ImageOutput"]; - flux_lora_collection_loader: components["schemas"]["FluxLoRALoaderOutput"]; - flux_lora_loader: components["schemas"]["FluxLoRALoaderOutput"]; - flux_model_loader: components["schemas"]["FluxModelLoaderOutput"]; - flux_model_loader_input: components["schemas"]["FluxModelLoaderOutput"]; - flux_model_to_string: components["schemas"]["StringOutput"]; - flux_redux: components["schemas"]["FluxReduxOutput"]; - flux_redux_collection: components["schemas"]["FluxReduxCollectionOutput"]; - flux_redux_collection_join: components["schemas"]["FluxReduxCollectionOutput"]; - flux_redux_conditioning_math: components["schemas"]["FluxReduxOutput"]; - flux_redux_downsampling: components["schemas"]["FluxReduxOutput"]; - flux_redux_index: components["schemas"]["FluxReduxOutput"]; - flux_redux_linked: components["schemas"]["FluxReduxListOutput"]; - flux_redux_rescale_conditioning: components["schemas"]["FluxReduxOutput"]; - flux_rescale_conditioning: components["schemas"]["FluxConditioningOutput"]; - flux_scale_conditioning: components["schemas"]["FluxConditioningOutput"]; - flux_scale_prompt_section: components["schemas"]["FluxConditioningOutput"]; - flux_scale_redux_conditioning: components["schemas"]["FluxReduxOutput"]; - flux_text_encoder: components["schemas"]["FluxConditioningOutput"]; - flux_text_encoder_linked: components["schemas"]["FluxTextEncoderListOutput"]; - flux_vae_decode: components["schemas"]["ImageOutput"]; - flux_vae_encode: components["schemas"]["LatentsOutput"]; - flux_weighted_prompt: components["schemas"]["FluxWeightedPromptOutput"]; - freeu: components["schemas"]["UNetOutput"]; - frequency_blend_latents: components["schemas"]["LatentsOutput"]; - frequency_match_latents: components["schemas"]["LatentsOutput"]; - generate_evolutionary_prompts: components["schemas"]["EvolutionaryPromptListOutput"]; - get_image_mask_bounding_box: components["schemas"]["BoundingBoxOutput"]; - get_palette: components["schemas"]["ImageOutput"]; - get_palette_adv: components["schemas"]["ImageOutput"]; - get_source_framerate: components["schemas"]["FloatOutput"]; - get_total_frames: components["schemas"]["IntegerOutput"]; - grounding_dino: components["schemas"]["BoundingBoxCollectionOutput"]; - halftone: components["schemas"]["ImageOutput"]; - hed_edge_detection: components["schemas"]["ImageOutput"]; - heuristic_resize: components["schemas"]["ImageOutput"]; - i2l: components["schemas"]["LatentsOutput"]; - ideal_size: components["schemas"]["IdealSizeOutput"]; - image: components["schemas"]["ImageOutput"]; - image_batch: components["schemas"]["ImageOutput"]; - image_collection: components["schemas"]["ImageCollectionOutput"]; - image_collection_index: components["schemas"]["ImageOutput"]; - image_collection_linked: components["schemas"]["ImageCollectionOutput"]; - image_collection_toggle: components["schemas"]["ImageCollectionOutput"]; - image_generator: components["schemas"]["ImageGeneratorOutput"]; - image_index_collect: components["schemas"]["ImageIndexCollectOutput"]; - image_mask_to_tensor: components["schemas"]["MaskOutput"]; - image_panel_layout: components["schemas"]["ImagePanelCoordinateOutput"]; - image_som: components["schemas"]["ImageSOMOutput"]; - image_to_image_name: components["schemas"]["ImageToImageNameOutput"]; - image_to_xy_image_collection: components["schemas"]["StringCollectionOutput"]; - image_to_xy_image_tiles: components["schemas"]["ImageToXYImageTilesOutput"]; - image_toggle: components["schemas"]["ImageOutput"]; - images_index_to_video: components["schemas"]["ImagesIndexToVideoOutput"]; - images_to_grids: components["schemas"]["ImageCollectionOutput"]; - img_blur: components["schemas"]["ImageOutput"]; - img_chan: components["schemas"]["ImageOutput"]; - img_channel_multiply: components["schemas"]["ImageOutput"]; - img_channel_offset: components["schemas"]["ImageOutput"]; - img_conv: components["schemas"]["ImageOutput"]; - img_crop: components["schemas"]["ImageOutput"]; - img_hue_adjust: components["schemas"]["ImageOutput"]; - img_ilerp: components["schemas"]["ImageOutput"]; - img_lerp: components["schemas"]["ImageOutput"]; - img_mul: components["schemas"]["ImageOutput"]; - img_noise: components["schemas"]["ImageOutput"]; - img_nsfw: components["schemas"]["ImageOutput"]; - img_pad_crop: components["schemas"]["ImageOutput"]; - img_paste: components["schemas"]["ImageOutput"]; - img_resize: components["schemas"]["ImageOutput"]; - img_scale: components["schemas"]["ImageOutput"]; - img_watermark: components["schemas"]["ImageOutput"]; - imgs2mask_clipseg: components["schemas"]["ImageOutput"]; - infill_cv2: components["schemas"]["ImageOutput"]; - infill_lama: components["schemas"]["ImageOutput"]; - infill_patchmatch: components["schemas"]["ImageOutput"]; - infill_rgba: components["schemas"]["ImageOutput"]; - infill_tile: components["schemas"]["ImageOutput"]; - info_grabber_unet_invocation: components["schemas"]["StringOutput"]; - int_collection_toggle: components["schemas"]["IntegerCollectionOutput"]; - int_toggle: components["schemas"]["IntegerOutput"]; - integer: components["schemas"]["IntegerOutput"]; - integer_batch: components["schemas"]["IntegerOutput"]; - integer_collection: components["schemas"]["IntegerCollectionOutput"]; - integer_collection_index: components["schemas"]["IntegerOutput"]; - integer_collection_linked: components["schemas"]["IntegerCollectionOutput"]; - integer_generator: components["schemas"]["IntegerGeneratorOutput"]; - integer_math: components["schemas"]["IntegerOutput"]; - ints_to_strings: components["schemas"]["StringCollectionOutput"]; - invert_tensor_mask: components["schemas"]["MaskOutput"]; - invokeai_ealightness: components["schemas"]["ImageOutput"]; - invokeai_img_blend: components["schemas"]["ImageOutput"]; - invokeai_img_composite: components["schemas"]["ImageOutput"]; - invokeai_img_dilate_erode: components["schemas"]["ImageOutput"]; - invokeai_img_enhance: components["schemas"]["ImageOutput"]; - invokeai_img_hue_adjust_plus: components["schemas"]["ImageOutput"]; - invokeai_img_val_thresholds: components["schemas"]["ImageOutput"]; - ip_adapter: components["schemas"]["IPAdapterOutput"]; - ip_adapter_linked: components["schemas"]["IPAdapterListOutput"]; - iterate: components["schemas"]["IterateInvocationOutput"]; - l2i: components["schemas"]["ImageOutput"]; - latent_average: components["schemas"]["LatentsOutput"]; - latent_band_pass_filter: components["schemas"]["LatentsOutput"]; - latent_blend_linear: components["schemas"]["LatentsOutput"]; - latent_channels_to_grid: components["schemas"]["ImageCollectionOutput"]; - latent_combine: components["schemas"]["LatentsOutput"]; - latent_dtype_convert: components["schemas"]["LatentsOutput"]; - latent_high_pass_filter: components["schemas"]["LatentsOutput"]; - latent_low_pass_filter: components["schemas"]["LatentsOutput"]; - latent_match: components["schemas"]["LatentsOutput"]; - latent_modify_channels: components["schemas"]["LatentsOutput"]; - latent_normalize_range: components["schemas"]["LatentsOutput"]; - latent_normalize_std_dev: components["schemas"]["LatentsOutput"]; - latent_normalize_std_dev_range: components["schemas"]["LatentsOutput"]; - latent_plot: components["schemas"]["ImageCollectionOutput"]; - latent_som: components["schemas"]["LatentsOutput"]; - latent_white_noise: components["schemas"]["LatentsOutput"]; - latents: components["schemas"]["LatentsOutput"]; - latents_collection: components["schemas"]["LatentsCollectionOutput"]; - latents_collection_index: components["schemas"]["LatentsOutput"]; - latents_collection_linked: components["schemas"]["LatentsCollectionOutput"]; - lblend: components["schemas"]["LatentsOutput"]; - lcrop: components["schemas"]["LatentsOutput"]; - lens_aperture_generator: components["schemas"]["ImageOutput"]; - lens_blur: components["schemas"]["ImageOutput"]; - lens_vignette: components["schemas"]["ImageOutput"]; - lineart_anime_edge_detection: components["schemas"]["ImageOutput"]; - lineart_edge_detection: components["schemas"]["ImageOutput"]; - llava_onevision_vllm: components["schemas"]["StringOutput"]; - load_all_text_files_in_folder_output: components["schemas"]["LoadAllTextFilesInFolderOutput"]; - load_aperture_image: components["schemas"]["ImageOutput"]; - load_text_file_to_string_invocation: components["schemas"]["LoadTextFileToStringOutput"]; - load_video_frame: components["schemas"]["ImageOutput"]; - lookup_from_prompt: components["schemas"]["LookupTableOutput"]; - lookup_lora_collection_triggers_invocation: components["schemas"]["LookupLoRATriggersOutput"]; - lookup_lora_triggers_invocation: components["schemas"]["LookupLoRATriggersOutput"]; - lookup_table_from_file: components["schemas"]["LookupTableOutput"]; - lora_collection: components["schemas"]["LoRACollectionOutput"]; - lora_collection_from_path_invocation: components["schemas"]["LoRACollectionFromPathOutput"]; - lora_collection_linked: components["schemas"]["LoRACollectionOutput"]; - lora_collection_loader: components["schemas"]["LoRALoaderOutput"]; - lora_collection_toggle: components["schemas"]["LoRACollectionToggleOutput"]; - lora_loader: components["schemas"]["LoRALoaderOutput"]; - lora_name_grabber_invocation: components["schemas"]["LoRANameGrabberOutput"]; - lora_selector: components["schemas"]["LoRASelectorOutput"]; - lora_to_string: components["schemas"]["StringOutput"]; - lora_toggle: components["schemas"]["LoRAToggleOutput"]; - lresize: components["schemas"]["LatentsOutput"]; - lscale: components["schemas"]["LatentsOutput"]; - main_model_loader: components["schemas"]["ModelLoaderOutput"]; - main_model_loader_input: components["schemas"]["ModelLoaderOutput"]; - main_model_to_string: components["schemas"]["StringOutput"]; - mask_combine: components["schemas"]["ImageOutput"]; - mask_edge: components["schemas"]["ImageOutput"]; - mask_from_id: components["schemas"]["ImageOutput"]; - match_histogram: components["schemas"]["ImageOutput"]; - match_histogram_lab: components["schemas"]["ImageOutput"]; - math_eval: components["schemas"]["MathEvalOutput"]; - mediapipe_face_detection: components["schemas"]["ImageOutput"]; - merge_lora_collections_invocation: components["schemas"]["MergeLoRACollectionsOutput"]; - merge_metadata: components["schemas"]["MetadataOutput"]; - merge_string_collections_invocation: components["schemas"]["MergeStringCollectionsOutput"]; - merge_tiles_to_image: components["schemas"]["ImageOutput"]; - metadata: components["schemas"]["MetadataOutput"]; - metadata_field_extractor: components["schemas"]["StringOutput"]; - metadata_from_image: components["schemas"]["MetadataOutput"]; - metadata_item: components["schemas"]["MetadataItemOutput"]; - metadata_item_linked: components["schemas"]["MetadataOutput"]; - metadata_to_bool: components["schemas"]["BooleanOutput"]; - metadata_to_bool_collection: components["schemas"]["BooleanCollectionOutput"]; - metadata_to_controlnets: components["schemas"]["MDControlListOutput"]; - metadata_to_float: components["schemas"]["FloatOutput"]; - metadata_to_float_collection: components["schemas"]["FloatCollectionOutput"]; - metadata_to_integer: components["schemas"]["IntegerOutput"]; - metadata_to_integer_collection: components["schemas"]["IntegerCollectionOutput"]; - metadata_to_ip_adapters: components["schemas"]["MDIPAdapterListOutput"]; - metadata_to_lora_collection: components["schemas"]["MetadataToLorasCollectionOutput"]; - metadata_to_loras: components["schemas"]["LoRALoaderOutput"]; - metadata_to_model: components["schemas"]["MetadataToModelOutput"]; - metadata_to_scheduler: components["schemas"]["SchedulerOutput"]; - metadata_to_sdlx_loras: components["schemas"]["SDXLLoRALoaderOutput"]; - metadata_to_sdxl_model: components["schemas"]["MetadataToSDXLModelOutput"]; - metadata_to_string: components["schemas"]["StringOutput"]; - metadata_to_string_collection: components["schemas"]["StringCollectionOutput"]; - metadata_to_t2i_adapters: components["schemas"]["MDT2IAdapterListOutput"]; - metadata_to_vae: components["schemas"]["VAEOutput"]; - minimum_overlap_xy_tile_generator: components["schemas"]["TilesOutput"]; - mlsd_detection: components["schemas"]["ImageOutput"]; - model_identifier: components["schemas"]["ModelIdentifierOutput"]; - model_name_grabber_invocation: components["schemas"]["ModelNameGrabberOutput"]; - model_name_to_model: components["schemas"]["StringToModelOutput"]; - model_to_string: components["schemas"]["StringOutput"]; - model_toggle: components["schemas"]["ModelToggleOutput"]; - monochrome_film_grain: components["schemas"]["ImageOutput"]; - mul: components["schemas"]["IntegerOutput"]; - nightmare_promptgen: components["schemas"]["NightmareOutput"]; - noise: components["schemas"]["NoiseOutput"]; - noise_add_flux: components["schemas"]["LatentsOutput"]; - noise_coordinated: components["schemas"]["NoiseOutput"]; - noise_coordinated_flux: components["schemas"]["LatentsOutput"]; - noise_spectral: components["schemas"]["NoiseOutput"]; - noiseimg_2d: components["schemas"]["ImageOutput"]; - normal_map: components["schemas"]["ImageOutput"]; - octree_quantizer: components["schemas"]["ImageOutput"]; - offset_image: components["schemas"]["ImageOutput"]; - offset_latents: components["schemas"]["LatentsOutput"]; - optimized_tile_size_from_area: components["schemas"]["TileSizeOutput"]; - pair_tile_image: components["schemas"]["PairTileImageOutput"]; - parse_weighted_string: components["schemas"]["WeightedStringOutput"]; - paste_image_into_bounding_box: components["schemas"]["ImageOutput"]; - pbr_maps: components["schemas"]["PBRMapsOutput"]; - percent_to_float: components["schemas"]["FloatOutput"]; - pidi_edge_detection: components["schemas"]["ImageOutput"]; - "pixel-art": components["schemas"]["ImageOutput"]; - pixelize: components["schemas"]["ImageOutput"]; - print_string_to_console_invocation: components["schemas"]["PrintStringToConsoleOutput"]; - prompt_auto_and: components["schemas"]["StringOutput"]; - prompt_from_file: components["schemas"]["StringCollectionOutput"]; - prompt_from_lookup_table: components["schemas"]["HalvedPromptOutput"]; - prompt_strength: components["schemas"]["PromptStrengthOutput"]; - prompt_strengths_combine: components["schemas"]["StringOutput"]; - prompt_template: components["schemas"]["PromptTemplateOutput"]; - prompt_to_file: components["schemas"]["PromptsToFileInvocationOutput"]; - prune_prompt: components["schemas"]["PrunedPromptOutput"]; - pt_fields_collect: components["schemas"]["PTFieldsCollectOutput"]; - pt_fields_expand: components["schemas"]["PTFieldsExpandOutput"]; - quote_escaper: components["schemas"]["EscapedOutput"]; - qwen3_prompt_pro: components["schemas"]["Qwen3PromptProOutput"]; - rand_float: components["schemas"]["FloatOutput"]; - rand_int: components["schemas"]["IntegerOutput"]; - random_image_size_invocation: components["schemas"]["RandomImageSizeOutput"]; - random_lora_mixer_invocation: components["schemas"]["RandomLoRAMixerOutput"]; - random_range: components["schemas"]["IntegerCollectionOutput"]; - range: components["schemas"]["IntegerCollectionOutput"]; - range_of_size: components["schemas"]["IntegerCollectionOutput"]; - reapply_lora_weight_invocation: components["schemas"]["ReapplyLoRAWeightOutput"]; - rectangle_mask: components["schemas"]["MaskOutput"]; - retrieve_flux_conditioning: components["schemas"]["RetrieveFluxConditioningMultiOutput"]; - retro_bitize: components["schemas"]["ImageOutput"]; - retro_crt_curvature: components["schemas"]["ImageOutput"]; - retro_palettize: components["schemas"]["ImageOutput"]; - retro_palettize_adv: components["schemas"]["ImageOutput"]; - retro_pixelize: components["schemas"]["ImageOutput"]; - retro_quantize: components["schemas"]["ImageOutput"]; - retro_scanlines_simple: components["schemas"]["ImageOutput"]; - rgb_merge: components["schemas"]["ImageOutput"]; - rgb_split: components["schemas"]["RGBSplitOutput"]; - rotate_image: components["schemas"]["ImageOutput"]; - round_float: components["schemas"]["FloatOutput"]; - save_image: components["schemas"]["ImageOutput"]; - scheduler: components["schemas"]["SchedulerOutput"]; - scheduler_to_string: components["schemas"]["StringOutput"]; - scheduler_toggle: components["schemas"]["SchedulerOutput"]; - sd3_denoise: components["schemas"]["LatentsOutput"]; - sd3_i2l: components["schemas"]["LatentsOutput"]; - sd3_l2i: components["schemas"]["ImageOutput"]; - sd3_model_loader: components["schemas"]["Sd3ModelLoaderOutput"]; - sd3_model_loader_input: components["schemas"]["Sd3ModelLoaderOutput"]; - sd3_model_to_string: components["schemas"]["StringOutput"]; - sd3_text_encoder: components["schemas"]["SD3ConditioningOutput"]; - sdxl_compel_prompt: components["schemas"]["ConditioningOutput"]; - sdxl_lora_collection_loader: components["schemas"]["SDXLLoRALoaderOutput"]; - sdxl_lora_loader: components["schemas"]["SDXLLoRALoaderOutput"]; - sdxl_main_model_toggle: components["schemas"]["SDXLModelLoaderOutput"]; - sdxl_model_loader: components["schemas"]["SDXLModelLoaderOutput"]; - sdxl_model_loader_input: components["schemas"]["SDXLModelLoaderOutput"]; - sdxl_model_to_string: components["schemas"]["StringOutput"]; - sdxl_refiner_compel_prompt: components["schemas"]["ConditioningOutput"]; - sdxl_refiner_model_loader: components["schemas"]["SDXLRefinerModelLoaderOutput"]; - seamless: components["schemas"]["SeamlessModeOutput"]; - segment_anything: components["schemas"]["MaskOutput"]; - separate_prompt_and_seed_vector: components["schemas"]["JsonListStringsOutput"]; - shmmask: components["schemas"]["ShadowsHighlightsMidtonesMasksOutput"]; - show_image: components["schemas"]["ImageOutput"]; - spandrel_image_to_image: components["schemas"]["ImageOutput"]; - spandrel_image_to_image_autoscale: components["schemas"]["ImageOutput"]; - spherical_distortion: components["schemas"]["ImageOutput"]; - store_flux_conditioning: components["schemas"]["FluxConditioningStoreOutput"]; - string: components["schemas"]["StringOutput"]; - string_batch: components["schemas"]["StringOutput"]; - string_collection: components["schemas"]["StringCollectionOutput"]; - string_collection_index: components["schemas"]["StringOutput"]; - string_collection_joiner_invocation: components["schemas"]["StringCollectionJoinerOutput"]; - string_collection_linked: components["schemas"]["StringCollectionOutput"]; - string_collection_toggle: components["schemas"]["StringCollectionOutput"]; - string_generator: components["schemas"]["StringGeneratorOutput"]; - string_join: components["schemas"]["StringOutput"]; - string_join_three: components["schemas"]["StringOutput"]; - string_replace: components["schemas"]["StringOutput"]; - string_split: components["schemas"]["String2Output"]; - string_split_neg: components["schemas"]["StringPosNegOutput"]; - string_to_collection_splitter_invocation: components["schemas"]["StringCollectionOutput"]; - string_to_float: components["schemas"]["FloatOutput"]; - string_to_int: components["schemas"]["IntegerOutput"]; - string_to_lora: components["schemas"]["StringToLoraOutput"]; - string_to_main_model: components["schemas"]["StringToMainModelOutput"]; - string_to_model: components["schemas"]["StringToModelOutput"]; - string_to_scheduler: components["schemas"]["SchedulerOutput"]; - string_to_sdxl_model: components["schemas"]["StringToSDXLModelOutput"]; - string_toggle: components["schemas"]["StringOutput"]; - strings_to_csv: components["schemas"]["StringOutput"]; - sub: components["schemas"]["IntegerOutput"]; - t2i_adapter: components["schemas"]["T2IAdapterOutput"]; - t2i_adapter_linked: components["schemas"]["T2IAdapterListOutput"]; - tensor_mask_to_image: components["schemas"]["ImageOutput"]; - text_mask: components["schemas"]["ImageOutput"]; - thresholding: components["schemas"]["ThresholdingOutput"]; - tile_to_properties: components["schemas"]["TileToPropertiesOutput"]; - tiled_multi_diffusion_denoise_latents: components["schemas"]["LatentsOutput"]; - tomask: components["schemas"]["ImageOutput"]; - tracery_invocation: components["schemas"]["TraceryOutput"]; - txt2mask_clipseg: components["schemas"]["ImageOutput"]; - txt2mask_clipseg_adv: components["schemas"]["ImageOutput"]; - unsharp_mask: components["schemas"]["ImageOutput"]; - vae_loader: components["schemas"]["VAEOutput"]; - xy_expand: components["schemas"]["XYExpandOutput"]; - xy_image_collect: components["schemas"]["StringOutput"]; - xy_image_expand: components["schemas"]["XYImageExpandOutput"]; - xy_image_tiles_to_image: components["schemas"]["ImageOutput"]; - xy_images_to_grid: components["schemas"]["ImageOutput"]; - xy_product: components["schemas"]["XYProductOutput"]; - xy_product_csv: components["schemas"]["XYProductOutput"]; - z_image_control: components["schemas"]["ZImageControlOutput"]; - z_image_denoise: components["schemas"]["LatentsOutput"]; - z_image_denoise_meta: components["schemas"]["LatentsMetaOutput"]; - z_image_i2l: components["schemas"]["LatentsOutput"]; - z_image_l2i: components["schemas"]["ImageOutput"]; - z_image_lora_collection_loader: components["schemas"]["ZImageLoRALoaderOutput"]; - z_image_lora_loader: components["schemas"]["ZImageLoRALoaderOutput"]; - z_image_model_loader: components["schemas"]["ZImageModelLoaderOutput"]; - z_image_seed_variance_enhancer: components["schemas"]["ZImageConditioningOutput"]; - z_image_text_encoder: components["schemas"]["ZImageConditioningOutput"]; - }; - /** - * InvocationProgressEvent - * @description Event model for invocation_progress - */ - InvocationProgressEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; - /** - * Queue Id - * @description The ID of the queue - */ - queue_id: string; - /** - * Item Id - * @description The ID of the queue item - */ - item_id: number; - /** - * Batch Id - * @description The ID of the queue batch - */ - batch_id: string; - /** - * Origin - * @description The origin of the queue item - * @default null - */ - origin: string | null; - /** - * Destination - * @description The destination of the queue item - * @default null - */ - destination: string | null; - /** - * User Id - * @description The ID of the user who created the queue item - * @default system - */ - user_id: string; - /** - * Session Id - * @description The ID of the session (aka graph execution state) - */ - session_id: string; - /** - * Invocation - * @description The ID of the invocation - */ - invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; - /** - * Invocation Source Id - * @description The ID of the prepared invocation's source node - */ - invocation_source_id: string; - /** - * Message - * @description A message to display - */ - message: string; - /** - * Percentage - * @description The percentage of the progress (omit to indicate indeterminate progress) - * @default null - */ - percentage: number | null; - /** - * @description An image representing the current state of the progress - * @default null - */ - image: components["schemas"]["ProgressImage"] | null; - }; - /** - * InvocationStartedEvent - * @description Event model for invocation_started - */ - InvocationStartedEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; - /** - * Queue Id - * @description The ID of the queue - */ - queue_id: string; - /** - * Item Id - * @description The ID of the queue item - */ - item_id: number; - /** - * Batch Id - * @description The ID of the queue batch - */ - batch_id: string; - /** - * Origin - * @description The origin of the queue item - * @default null - */ - origin: string | null; - /** - * Destination - * @description The destination of the queue item - * @default null - */ - destination: string | null; - /** - * User Id - * @description The ID of the user who created the queue item - * @default system - */ - user_id: string; - /** - * Session Id - * @description The ID of the session (aka graph execution state) - */ - session_id: string; - /** - * Invocation - * @description The ID of the invocation - */ - invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AdvAutostereogramInvocation"] | components["schemas"]["AdvancedTextFontImageInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["AnamorphicStreaksInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["AutostereogramInvocation"] | components["schemas"]["AverageImagesInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BoolCollectionIndexInvocation"] | components["schemas"]["BoolCollectionToggleInvocation"] | components["schemas"]["BoolToggleInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanCollectionLinkedInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CMYKColorSeparationInvocation"] | components["schemas"]["CMYKHalftoneInvocation"] | components["schemas"]["CMYKMergeInvocation"] | components["schemas"]["CMYKSplitInvocation"] | components["schemas"]["CSVToIndexStringInvocation"] | components["schemas"]["CSVToStringsInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["ChromaNoiseInvocation"] | components["schemas"]["ChromaticAberrationInvocation"] | components["schemas"]["ClipsegMaskHierarchyInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["CollectionCountInvocation"] | components["schemas"]["CollectionIndexInvocation"] | components["schemas"]["CollectionJoinInvocation"] | components["schemas"]["CollectionReverseInvocation"] | components["schemas"]["CollectionSliceInvocation"] | components["schemas"]["CollectionSortInvocation"] | components["schemas"]["CollectionUniqueInvocation"] | components["schemas"]["ColorCastCorrectionInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompareFloatsInvocation"] | components["schemas"]["CompareIntsInvocation"] | components["schemas"]["CompareStringsInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConcatenateFluxConditioningInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningCollectionLinkedInvocation"] | components["schemas"]["ConditioningCollectionToggleInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ConditioningToggleInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["ControlNetLinkedInvocation"] | components["schemas"]["CoordinatedFluxNoiseInvocation"] | components["schemas"]["CoordinatedNoiseInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CropLatentsInvocation"] | components["schemas"]["CrossoverPromptInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DefaultXYTileGenerator"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["EnhanceDetailInvocation"] | components["schemas"]["EscaperInvocation"] | components["schemas"]["EvenSplitXYTileGenerator"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["ExtractImageCollectionMetadataBooleanInvocation"] | components["schemas"]["ExtractImageCollectionMetadataFloatInvocation"] | components["schemas"]["ExtractImageCollectionMetadataIntegerInvocation"] | components["schemas"]["ExtractImageCollectionMetadataStringInvocation"] | components["schemas"]["ExtrudeDepthInvocation"] | components["schemas"]["FLUXConditioningCollectionToggleInvocation"] | components["schemas"]["FLUXConditioningToggleInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FilmGrainInvocation"] | components["schemas"]["FlattenHistogramMono"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionIndexInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatCollectionLinkedInvocation"] | components["schemas"]["FloatCollectionToggleInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["FloatToggleInvocation"] | components["schemas"]["FloatsToStringsInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxConditioningBlendInvocation"] | components["schemas"]["FluxConditioningCollectionIndexInvocation"] | components["schemas"]["FluxConditioningCollectionInvocation"] | components["schemas"]["FluxConditioningCollectionJoinInvocation"] | components["schemas"]["FluxConditioningDeltaAugmentationInvocation"] | components["schemas"]["FluxConditioningListInvocation"] | components["schemas"]["FluxConditioningMathOperationInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetCollectionIndexInvocation"] | components["schemas"]["FluxControlNetCollectionInvocation"] | components["schemas"]["FluxControlNetCollectionJoinInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxControlNetLinkedInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxIdealSizeInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextIdealSizeInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInputInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxModelToStringInvocation"] | components["schemas"]["FluxReduxCollectionIndexInvocation"] | components["schemas"]["FluxReduxCollectionInvocation"] | components["schemas"]["FluxReduxCollectionJoinInvocation"] | components["schemas"]["FluxReduxConditioningMathOperationInvocation"] | components["schemas"]["FluxReduxDownsamplingInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxReduxLinkedInvocation"] | components["schemas"]["FluxReduxRescaleConditioningInvocation"] | components["schemas"]["FluxRescaleConditioningInvocation"] | components["schemas"]["FluxScaleConditioningInvocation"] | components["schemas"]["FluxScalePromptSectionInvocation"] | components["schemas"]["FluxScaleReduxConditioningInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxTextEncoderLinkedInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FluxWeightedPromptInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["FrequencyBlendLatents"] | components["schemas"]["FrequencySpectrumMatchLatentsInvocation"] | components["schemas"]["GenerateEvolutionaryPromptsInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GetSourceFrameRateInvocation"] | components["schemas"]["GetTotalFramesInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HalftoneInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IPAdapterLinkedInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionIndexInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageCollectionLinkedInvocation"] | components["schemas"]["ImageCollectionToggleInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageIndexCollectInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImageOffsetInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageRotateInvocation"] | components["schemas"]["ImageSOMInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageSearchToMaskClipsegInvocation"] | components["schemas"]["ImageToAAInvocation"] | components["schemas"]["ImageToDetailedASCIIArtInvocation"] | components["schemas"]["ImageToImageNameInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageToUnicodeArtInvocation"] | components["schemas"]["ImageToXYImageCollectionInvocation"] | components["schemas"]["ImageToXYImageTilesInvocation"] | components["schemas"]["ImageToggleInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["ImagesIndexToVideoInvocation"] | components["schemas"]["ImagesToGridsInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["InfoGrabberUNetInvocation"] | components["schemas"]["IntCollectionToggleInvocation"] | components["schemas"]["IntToggleInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionIndexInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerCollectionLinkedInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["IntsToStringsInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentAverageInvocation"] | components["schemas"]["LatentBandPassFilterInvocation"] | components["schemas"]["LatentBlendLinearInvocation"] | components["schemas"]["LatentChannelsToGridInvocation"] | components["schemas"]["LatentCombineInvocation"] | components["schemas"]["LatentDtypeConvertInvocation"] | components["schemas"]["LatentHighPassFilterInvocation"] | components["schemas"]["LatentLowPassFilterInvocation"] | components["schemas"]["LatentMatchInvocation"] | components["schemas"]["LatentModifyChannelsInvocation"] | components["schemas"]["LatentNormalizeRangeInvocation"] | components["schemas"]["LatentNormalizeStdDevInvocation"] | components["schemas"]["LatentNormalizeStdRangeInvocation"] | components["schemas"]["LatentPlotInvocation"] | components["schemas"]["LatentSOMInvocation"] | components["schemas"]["LatentWhiteNoiseInvocation"] | components["schemas"]["LatentsCollectionIndexInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsCollectionLinkedInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LensApertureGeneratorInvocation"] | components["schemas"]["LensBlurInvocation"] | components["schemas"]["LensVignetteInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionFromPathInvocation"] | components["schemas"]["LoRACollectionInvocation"] | components["schemas"]["LoRACollectionLinkedInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRACollectionToggleInvocation"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRANameGrabberInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["LoRAToggleInvocation"] | components["schemas"]["LoadAllTextFilesInFolderInvocation"] | components["schemas"]["LoadApertureImageInvocation"] | components["schemas"]["LoadTextFileToStringInvocation"] | components["schemas"]["LoadVideoFrameInvocation"] | components["schemas"]["LookupLoRACollectionTriggersInvocation"] | components["schemas"]["LookupLoRATriggersInvocation"] | components["schemas"]["LookupTableFromFileInvocation"] | components["schemas"]["LookupsEntryFromPromptInvocation"] | components["schemas"]["LoraToStringInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInputInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MainModelToStringInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MatchHistogramInvocation"] | components["schemas"]["MatchHistogramLabInvocation"] | components["schemas"]["MathEvalInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeLoRACollectionsInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeStringCollectionsInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["MinimumOverlapXYTileGenerator"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["ModelNameGrabberInvocation"] | components["schemas"]["ModelNameToModelInvocation"] | components["schemas"]["ModelToStringInvocation"] | components["schemas"]["ModelToggleInvocation"] | components["schemas"]["MonochromeFilmGrainInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NightmareInvocation"] | components["schemas"]["NoiseAddFluxInvocation"] | components["schemas"]["NoiseImage2DInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NoiseSpectralInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["OctreeQuantizerInvocation"] | components["schemas"]["OffsetLatentsInvocation"] | components["schemas"]["OptimizedTileSizeFromAreaInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PTFieldsCollectInvocation"] | components["schemas"]["PTFieldsExpandInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["ParseWeightedStringInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PercentToFloatInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PixelArtInvocation"] | components["schemas"]["PixelizeImageInvocation"] | components["schemas"]["PixelizeInvocation"] | components["schemas"]["PrintStringToConsoleInvocation"] | components["schemas"]["PromptAutoAndInvocation"] | components["schemas"]["PromptFromLookupTableInvocation"] | components["schemas"]["PromptStrengthInvocation"] | components["schemas"]["PromptStrengthsCombineInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["PromptsToFileInvocation"] | components["schemas"]["PruneTextInvocation"] | components["schemas"]["Qwen3PromptProInvocation"] | components["schemas"]["RGBMergeInvocation"] | components["schemas"]["RGBSplitInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomImageSizeInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomLoRAMixerInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["ReapplyLoRAWeightInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RetrieveBoardImagesInvocation"] | components["schemas"]["RetrieveFluxConditioningInvocation"] | components["schemas"]["RetroBitizeInvocation"] | components["schemas"]["RetroCRTCurvatureInvocation"] | components["schemas"]["RetroGetPaletteAdvInvocation"] | components["schemas"]["RetroGetPaletteInvocation"] | components["schemas"]["RetroPalettizeAdvInvocation"] | components["schemas"]["RetroPalettizeInvocation"] | components["schemas"]["RetroQuantizeInvocation"] | components["schemas"]["RetroScanlinesSimpleInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SD3ModelLoaderInputInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLMainModelToggleInvocation"] | components["schemas"]["SDXLModelLoaderInputInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLModelToStringInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["SchedulerToStringInvocation"] | components["schemas"]["SchedulerToggleInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3ModelToStringInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["SeparatePromptAndSeedVectorInvocation"] | components["schemas"]["ShadowsHighlightsMidtonesMaskInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["SphericalDistortionInvocation"] | components["schemas"]["StoreFluxConditioningInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionIndexInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringCollectionJoinerInvocation"] | components["schemas"]["StringCollectionLinkedInvocation"] | components["schemas"]["StringCollectionToggleInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["StringToCollectionSplitterInvocation"] | components["schemas"]["StringToFloatInvocation"] | components["schemas"]["StringToIntInvocation"] | components["schemas"]["StringToLoraInvocation"] | components["schemas"]["StringToMainModelInvocation"] | components["schemas"]["StringToModelInvocation"] | components["schemas"]["StringToSDXLModelInvocation"] | components["schemas"]["StringToSchedulerInvocation"] | components["schemas"]["StringToggleInvocation"] | components["schemas"]["StringsToCSVInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["T2IAdapterLinkedInvocation"] | components["schemas"]["TextMaskInvocation"] | components["schemas"]["TextToMaskClipsegAdvancedInvocation"] | components["schemas"]["TextToMaskClipsegInvocation"] | components["schemas"]["TextfontimageInvocation"] | components["schemas"]["ThresholdingInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["TraceryInvocation"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["XYExpandInvocation"] | components["schemas"]["XYImageCollectInvocation"] | components["schemas"]["XYImageExpandInvocation"] | components["schemas"]["XYImageTilesToImageInvocation"] | components["schemas"]["XYImagesToGridInvocation"] | components["schemas"]["XYProductCSVInvocation"] | components["schemas"]["XYProductInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; - /** - * Invocation Source Id - * @description The ID of the prepared invocation's source node - */ - invocation_source_id: string; - }; - /** - * InvokeAIAppConfig - * @description Invoke's global app configuration. - * - * Typically, you won't need to interact with this class directly. Instead, use the `get_config` function from `invokeai.app.services.config` to get a singleton config object. - * - * Attributes: - * host: IP address to bind to. Use `0.0.0.0` to serve to your local network. - * port: Port to bind to. - * allow_origins: Allowed CORS origins. - * allow_credentials: Allow CORS credentials. - * allow_methods: Methods allowed for CORS. - * allow_headers: Headers allowed for CORS. - * ssl_certfile: SSL certificate file for HTTPS. See https://www.uvicorn.org/settings/#https. - * ssl_keyfile: SSL key file for HTTPS. See https://www.uvicorn.org/settings/#https. - * log_tokenization: Enable logging of parsed prompt tokens. - * patchmatch: Enable patchmatch inpaint code. - * models_dir: Path to the models directory. - * convert_cache_dir: Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). - * download_cache_dir: Path to the directory that contains dynamically downloaded models. - * legacy_conf_dir: Path to directory of legacy checkpoint config files. - * db_dir: Path to InvokeAI databases directory. - * outputs_dir: Path to directory for outputs. - * custom_nodes_dir: Path to directory for custom nodes. - * style_presets_dir: Path to directory for style presets. - * workflow_thumbnails_dir: Path to directory for workflow thumbnails. - * log_handlers: Log handler. Valid options are "console", "file=", "syslog=path|address:host:port", "http=". - * log_format: Log format. Use "plain" for text-only, "color" for colorized output, "legacy" for 2.3-style logging and "syslog" for syslog-style.
Valid values: `plain`, `color`, `syslog`, `legacy` - * log_level: Emit logging messages at this level or higher.
Valid values: `debug`, `info`, `warning`, `error`, `critical` - * log_sql: Log SQL queries. `log_level` must be `debug` for this to do anything. Extremely verbose. - * log_level_network: Log level for network-related messages. 'info' and 'debug' are very verbose.
Valid values: `debug`, `info`, `warning`, `error`, `critical` - * use_memory_db: Use in-memory database. Useful for development. - * dev_reload: Automatically reload when Python sources are changed. Does not reload node definitions. - * profile_graphs: Enable graph profiling using `cProfile`. - * profile_prefix: An optional prefix for profile output files. - * profiles_dir: Path to profiles output directory. - * max_cache_ram_gb: The maximum amount of CPU RAM to use for model caching in GB. If unset, the limit will be configured based on the available RAM. In most cases, it is recommended to leave this unset. - * max_cache_vram_gb: The amount of VRAM to use for model caching in GB. If unset, the limit will be configured based on the available VRAM and the device_working_mem_gb. In most cases, it is recommended to leave this unset. - * log_memory_usage: If True, a memory snapshot will be captured before and after every model cache operation, and the result will be logged (at debug level). There is a time cost to capturing the memory snapshots, so it is recommended to only enable this feature if you are actively inspecting the model cache's behaviour. - * model_cache_keep_alive_min: How long to keep models in cache after last use, in minutes. A value of 0 (the default) means models are kept in cache indefinitely. If no model generations occur within the timeout period, the model cache is cleared using the same logic as the 'Clear Model Cache' button. - * device_working_mem_gb: The amount of working memory to keep available on the compute device (in GB). Has no effect if running on CPU. If you are experiencing OOM errors, try increasing this value. - * enable_partial_loading: Enable partial loading of models. This enables models to run with reduced VRAM requirements (at the cost of slower speed) by streaming the model from RAM to VRAM as its used. In some edge cases, partial loading can cause models to run more slowly if they were previously being fully loaded into VRAM. - * keep_ram_copy_of_weights: Whether to keep a full RAM copy of a model's weights when the model is loaded in VRAM. Keeping a RAM copy increases average RAM usage, but speeds up model switching and LoRA patching (assuming there is sufficient RAM). Set this to False if RAM pressure is consistently high. - * ram: DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_ram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. - * vram: DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_vram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. - * lazy_offload: DEPRECATED: This setting is no longer used. Lazy-offloading is enabled by default. This config setting will be removed once the new model cache behavior is stable. - * pytorch_cuda_alloc_conf: Configure the Torch CUDA memory allocator. This will impact peak reserved VRAM usage and performance. Setting to "backend:cudaMallocAsync" works well on many systems. The optimal configuration is highly dependent on the system configuration (device type, VRAM, CUDA driver version, etc.), so must be tuned experimentally. - * device: Preferred execution device. `auto` will choose the device depending on the hardware platform and the installed torch capabilities.
Valid values: `auto`, `cpu`, `cuda`, `mps`, `cuda:N` (where N is a device number) - * precision: Floating point precision. `float16` will consume half the memory of `float32` but produce slightly lower-quality images. The `auto` setting will guess the proper precision based on your video card and operating system.
Valid values: `auto`, `float16`, `bfloat16`, `float32` - * sequential_guidance: Whether to calculate guidance in serial instead of in parallel, lowering memory requirements. - * attention_type: Attention type.
Valid values: `auto`, `normal`, `xformers`, `sliced`, `torch-sdp` - * attention_slice_size: Slice size, valid when attention_type=="sliced".
Valid values: `auto`, `balanced`, `max`, `1`, `2`, `3`, `4`, `5`, `6`, `7`, `8` - * force_tiled_decode: Whether to enable tiled VAE decode (reduces memory consumption with some performance penalty). - * pil_compress_level: The compress_level setting of PIL.Image.save(), used for PNG encoding. All settings are lossless. 0 = no compression, 1 = fastest with slightly larger filesize, 9 = slowest with smallest filesize. 1 is typically the best setting. - * max_queue_size: Maximum number of items in the session queue. - * clear_queue_on_startup: Empties session queue on startup. - * allow_nodes: List of nodes to allow. Omit to allow all. - * deny_nodes: List of nodes to deny. Omit to deny none. - * node_cache_size: How many cached nodes to keep in memory. - * hashing_algorithm: Model hashing algorthim for model installs. 'blake3_multi' is best for SSDs. 'blake3_single' is best for spinning disk HDDs. 'random' disables hashing, instead assigning a UUID to models. Useful when using a memory db to reduce model installation time, or if you don't care about storing stable hashes for models. Alternatively, any other hashlib algorithm is accepted, though these are not nearly as performant as blake3.
Valid values: `blake3_multi`, `blake3_single`, `random`, `md5`, `sha1`, `sha224`, `sha256`, `sha384`, `sha512`, `blake2b`, `blake2s`, `sha3_224`, `sha3_256`, `sha3_384`, `sha3_512`, `shake_128`, `shake_256` - * remote_api_tokens: List of regular expression and token pairs used when downloading models from URLs. The download URL is tested against the regex, and if it matches, the token is provided in as a Bearer token. - * scan_models_on_startup: Scan the models directory on startup, registering orphaned models. This is typically only used in conjunction with `use_memory_db` for testing purposes. - * unsafe_disable_picklescan: UNSAFE. Disable the picklescan security check during model installation. Recommended only for development and testing purposes. This will allow arbitrary code execution during model installation, so should never be used in production. - * allow_unknown_models: Allow installation of models that we are unable to identify. If enabled, models will be marked as `unknown` in the database, and will not have any metadata associated with them. If disabled, unknown models will be rejected during installation. - * multiuser: Enable multiuser support. When disabled, the application runs in single-user mode using a default system account with administrator privileges. When enabled, requires user authentication and authorization. - * strict_password_checking: Enforce strict password requirements. When True, passwords must contain uppercase, lowercase, and numbers. When False (default), any password is accepted but its strength (weak/moderate/strong) is reported to the user. - */ - InvokeAIAppConfig: { - /** - * Schema Version - * @description Schema version of the config file. This is not a user-configurable setting. - * @default 4.0.2 - */ - schema_version?: string; - /** - * Legacy Models Yaml Path - * @description Path to the legacy models.yaml file. This is not a user-configurable setting. - */ - legacy_models_yaml_path?: string | null; - /** - * Host - * @description IP address to bind to. Use `0.0.0.0` to serve to your local network. - * @default 127.0.0.1 - */ - host?: string; - /** - * Port - * @description Port to bind to. - * @default 9090 - */ - port?: number; - /** - * Allow Origins - * @description Allowed CORS origins. - * @default [] - */ - allow_origins?: string[]; - /** - * Allow Credentials - * @description Allow CORS credentials. - * @default true - */ - allow_credentials?: boolean; - /** - * Allow Methods - * @description Methods allowed for CORS. - * @default [ - * "*" - * ] - */ - allow_methods?: string[]; - /** - * Allow Headers - * @description Headers allowed for CORS. - * @default [ - * "*" - * ] - */ - allow_headers?: string[]; - /** - * Ssl Certfile - * @description SSL certificate file for HTTPS. See https://www.uvicorn.org/settings/#https. - */ - ssl_certfile?: string | null; - /** - * Ssl Keyfile - * @description SSL key file for HTTPS. See https://www.uvicorn.org/settings/#https. - */ - ssl_keyfile?: string | null; - /** - * Log Tokenization - * @description Enable logging of parsed prompt tokens. - * @default false - */ - log_tokenization?: boolean; - /** - * Patchmatch - * @description Enable patchmatch inpaint code. - * @default true - */ - patchmatch?: boolean; - /** - * Models Dir - * Format: path - * @description Path to the models directory. - * @default models - */ - models_dir?: string; - /** - * Convert Cache Dir - * Format: path - * @description Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). - * @default models\.convert_cache - */ - convert_cache_dir?: string; - /** - * Download Cache Dir - * Format: path - * @description Path to the directory that contains dynamically downloaded models. - * @default models\.download_cache - */ - download_cache_dir?: string; - /** - * Legacy Conf Dir - * Format: path - * @description Path to directory of legacy checkpoint config files. - * @default configs - */ - legacy_conf_dir?: string; - /** - * Db Dir - * Format: path - * @description Path to InvokeAI databases directory. - * @default databases - */ - db_dir?: string; - /** - * Outputs Dir - * Format: path - * @description Path to directory for outputs. - * @default outputs - */ - outputs_dir?: string; - /** - * Custom Nodes Dir - * Format: path - * @description Path to directory for custom nodes. - * @default nodes - */ - custom_nodes_dir?: string; - /** - * Style Presets Dir - * Format: path - * @description Path to directory for style presets. - * @default style_presets - */ - style_presets_dir?: string; - /** - * Workflow Thumbnails Dir - * Format: path - * @description Path to directory for workflow thumbnails. - * @default workflow_thumbnails - */ - workflow_thumbnails_dir?: string; - /** - * Log Handlers - * @description Log handler. Valid options are "console", "file=", "syslog=path|address:host:port", "http=". - * @default [ - * "console" - * ] - */ - log_handlers?: string[]; - /** - * Log Format - * @description Log format. Use "plain" for text-only, "color" for colorized output, "legacy" for 2.3-style logging and "syslog" for syslog-style. - * @default color - * @enum {string} - */ - log_format?: "plain" | "color" | "syslog" | "legacy"; - /** - * Log Level - * @description Emit logging messages at this level or higher. - * @default info - * @enum {string} - */ - log_level?: "debug" | "info" | "warning" | "error" | "critical"; - /** - * Log Sql - * @description Log SQL queries. `log_level` must be `debug` for this to do anything. Extremely verbose. - * @default false - */ - log_sql?: boolean; - /** - * Log Level Network - * @description Log level for network-related messages. 'info' and 'debug' are very verbose. - * @default warning - * @enum {string} - */ - log_level_network?: "debug" | "info" | "warning" | "error" | "critical"; - /** - * Use Memory Db - * @description Use in-memory database. Useful for development. - * @default false - */ - use_memory_db?: boolean; - /** - * Dev Reload - * @description Automatically reload when Python sources are changed. Does not reload node definitions. - * @default false - */ - dev_reload?: boolean; - /** - * Profile Graphs - * @description Enable graph profiling using `cProfile`. - * @default false - */ - profile_graphs?: boolean; - /** - * Profile Prefix - * @description An optional prefix for profile output files. - */ - profile_prefix?: string | null; - /** - * Profiles Dir - * Format: path - * @description Path to profiles output directory. - * @default profiles - */ - profiles_dir?: string; - /** - * Max Cache Ram Gb - * @description The maximum amount of CPU RAM to use for model caching in GB. If unset, the limit will be configured based on the available RAM. In most cases, it is recommended to leave this unset. - */ - max_cache_ram_gb?: number | null; - /** - * Max Cache Vram Gb - * @description The amount of VRAM to use for model caching in GB. If unset, the limit will be configured based on the available VRAM and the device_working_mem_gb. In most cases, it is recommended to leave this unset. - */ - max_cache_vram_gb?: number | null; - /** - * Log Memory Usage - * @description If True, a memory snapshot will be captured before and after every model cache operation, and the result will be logged (at debug level). There is a time cost to capturing the memory snapshots, so it is recommended to only enable this feature if you are actively inspecting the model cache's behaviour. - * @default false - */ - log_memory_usage?: boolean; - /** - * Model Cache Keep Alive Min - * @description How long to keep models in cache after last use, in minutes. A value of 0 (the default) means models are kept in cache indefinitely. If no model generations occur within the timeout period, the model cache is cleared using the same logic as the 'Clear Model Cache' button. - * @default 0 - */ - model_cache_keep_alive_min?: number; - /** - * Device Working Mem Gb - * @description The amount of working memory to keep available on the compute device (in GB). Has no effect if running on CPU. If you are experiencing OOM errors, try increasing this value. - * @default 3 - */ - device_working_mem_gb?: number; - /** - * Enable Partial Loading - * @description Enable partial loading of models. This enables models to run with reduced VRAM requirements (at the cost of slower speed) by streaming the model from RAM to VRAM as its used. In some edge cases, partial loading can cause models to run more slowly if they were previously being fully loaded into VRAM. - * @default false - */ - enable_partial_loading?: boolean; - /** - * Keep Ram Copy Of Weights - * @description Whether to keep a full RAM copy of a model's weights when the model is loaded in VRAM. Keeping a RAM copy increases average RAM usage, but speeds up model switching and LoRA patching (assuming there is sufficient RAM). Set this to False if RAM pressure is consistently high. - * @default true - */ - keep_ram_copy_of_weights?: boolean; - /** - * Ram - * @description DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_ram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. - */ - ram?: number | null; - /** - * Vram - * @description DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_vram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. - */ - vram?: number | null; - /** - * Lazy Offload - * @description DEPRECATED: This setting is no longer used. Lazy-offloading is enabled by default. This config setting will be removed once the new model cache behavior is stable. - * @default true - */ - lazy_offload?: boolean; - /** - * Pytorch Cuda Alloc Conf - * @description Configure the Torch CUDA memory allocator. This will impact peak reserved VRAM usage and performance. Setting to "backend:cudaMallocAsync" works well on many systems. The optimal configuration is highly dependent on the system configuration (device type, VRAM, CUDA driver version, etc.), so must be tuned experimentally. - */ - pytorch_cuda_alloc_conf?: string | null; - /** - * Device - * @description Preferred execution device. `auto` will choose the device depending on the hardware platform and the installed torch capabilities.
Valid values: `auto`, `cpu`, `cuda`, `mps`, `cuda:N` (where N is a device number) - * @default auto - */ - device?: string; - /** - * Precision - * @description Floating point precision. `float16` will consume half the memory of `float32` but produce slightly lower-quality images. The `auto` setting will guess the proper precision based on your video card and operating system. - * @default auto - * @enum {string} - */ - precision?: "auto" | "float16" | "bfloat16" | "float32"; - /** - * Sequential Guidance - * @description Whether to calculate guidance in serial instead of in parallel, lowering memory requirements. - * @default false - */ - sequential_guidance?: boolean; - /** - * Attention Type - * @description Attention type. - * @default auto - * @enum {string} - */ - attention_type?: "auto" | "normal" | "xformers" | "sliced" | "torch-sdp"; - /** - * Attention Slice Size - * @description Slice size, valid when attention_type=="sliced". - * @default auto - * @enum {unknown} - */ - attention_slice_size?: "auto" | "balanced" | "max" | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8; - /** - * Force Tiled Decode - * @description Whether to enable tiled VAE decode (reduces memory consumption with some performance penalty). - * @default false - */ - force_tiled_decode?: boolean; - /** - * Pil Compress Level - * @description The compress_level setting of PIL.Image.save(), used for PNG encoding. All settings are lossless. 0 = no compression, 1 = fastest with slightly larger filesize, 9 = slowest with smallest filesize. 1 is typically the best setting. - * @default 1 - */ - pil_compress_level?: number; - /** - * Max Queue Size - * @description Maximum number of items in the session queue. - * @default 10000 - */ - max_queue_size?: number; - /** - * Clear Queue On Startup - * @description Empties session queue on startup. - * @default false - */ - clear_queue_on_startup?: boolean; - /** - * Allow Nodes - * @description List of nodes to allow. Omit to allow all. - */ - allow_nodes?: string[] | null; - /** - * Deny Nodes - * @description List of nodes to deny. Omit to deny none. - */ - deny_nodes?: string[] | null; - /** - * Node Cache Size - * @description How many cached nodes to keep in memory. - * @default 512 - */ - node_cache_size?: number; - /** - * Hashing Algorithm - * @description Model hashing algorthim for model installs. 'blake3_multi' is best for SSDs. 'blake3_single' is best for spinning disk HDDs. 'random' disables hashing, instead assigning a UUID to models. Useful when using a memory db to reduce model installation time, or if you don't care about storing stable hashes for models. Alternatively, any other hashlib algorithm is accepted, though these are not nearly as performant as blake3. - * @default blake3_single - * @enum {string} - */ - hashing_algorithm?: "blake3_multi" | "blake3_single" | "random" | "md5" | "sha1" | "sha224" | "sha256" | "sha384" | "sha512" | "blake2b" | "blake2s" | "sha3_224" | "sha3_256" | "sha3_384" | "sha3_512" | "shake_128" | "shake_256"; - /** - * Remote Api Tokens - * @description List of regular expression and token pairs used when downloading models from URLs. The download URL is tested against the regex, and if it matches, the token is provided in as a Bearer token. - */ - remote_api_tokens?: components["schemas"]["URLRegexTokenPair"][] | null; - /** - * Scan Models On Startup - * @description Scan the models directory on startup, registering orphaned models. This is typically only used in conjunction with `use_memory_db` for testing purposes. - * @default false - */ - scan_models_on_startup?: boolean; - /** - * Unsafe Disable Picklescan - * @description UNSAFE. Disable the picklescan security check during model installation. Recommended only for development and testing purposes. This will allow arbitrary code execution during model installation, so should never be used in production. - * @default false - */ - unsafe_disable_picklescan?: boolean; - /** - * Allow Unknown Models - * @description Allow installation of models that we are unable to identify. If enabled, models will be marked as `unknown` in the database, and will not have any metadata associated with them. If disabled, unknown models will be rejected during installation. - * @default true - */ - allow_unknown_models?: boolean; - /** - * Multiuser - * @description Enable multiuser support. When disabled, the application runs in single-user mode using a default system account with administrator privileges. When enabled, requires user authentication and authorization. - * @default false - */ - multiuser?: boolean; - /** - * Strict Password Checking - * @description Enforce strict password requirements. When True, passwords must contain uppercase, lowercase, and numbers. When False (default), any password is accepted but its strength (weak/moderate/strong) is reported to the user. - * @default false - */ - strict_password_checking?: boolean; - }; - /** - * InvokeAIAppConfigWithSetFields - * @description InvokeAI App Config with model fields set - */ - InvokeAIAppConfigWithSetFields: { - /** - * Set Fields - * @description The set fields - */ - set_fields: string[]; - /** @description The InvokeAI App Config */ - config: components["schemas"]["InvokeAIAppConfig"]; - }; - /** - * Adjust Image Hue Plus - * @description Adjusts the Hue of an image by rotating it in the selected color space. Originally created by @dwringer - */ - InvokeAdjustImageHuePlusInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to adjust - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Space - * @description Color space in which to rotate hue by polar coords (*: non-invertible) - * @default HSV / HSL / RGB - * @enum {string} - */ - space?: "HSV / HSL / RGB" | "Okhsl" | "Okhsv" | "*Oklch / Oklab" | "*LCh / CIELab" | "*UPLab (w/CIELab_to_UPLab.icc)"; - /** - * Degrees - * @description Degrees by which to rotate image hue - * @default 0 - */ - degrees?: number; - /** - * Preserve Lightness - * @description Whether to preserve CIELAB lightness values - * @default false - */ - preserve_lightness?: boolean; - /** - * Ok Adaptive Gamut - * @description Higher preserves chroma at the expense of lightness (Oklab) - * @default 0.05 - */ - ok_adaptive_gamut?: number; - /** - * Ok High Precision - * @description Use more steps in computing gamut (Oklab/Okhsv/Okhsl) - * @default true - */ - ok_high_precision?: boolean; - /** - * type - * @default invokeai_img_hue_adjust_plus - * @constant - */ - type: "invokeai_img_hue_adjust_plus"; - }; - /** - * Equivalent Achromatic Lightness - * @description Calculate Equivalent Achromatic Lightness from image. Originally created by @dwringer - */ - InvokeEquivalentAchromaticLightnessInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Image from which to get channel - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default invokeai_ealightness - * @constant - */ - type: "invokeai_ealightness"; - }; - /** - * Image Layer Blend - * @description Blend two images together, with optional opacity, mask, and blend modes. Originally created by @dwringer - */ - InvokeImageBlendInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The top image to blend - * @default null - */ - layer_upper?: components["schemas"]["ImageField"] | null; - /** - * Blend Mode - * @description Available blend modes - * @default Normal - * @enum {string} - */ - blend_mode?: "Normal" | "Lighten Only" | "Darken Only" | "Lighten Only (EAL)" | "Darken Only (EAL)" | "Hue" | "Saturation" | "Color" | "Luminosity" | "Linear Dodge (Add)" | "Subtract" | "Multiply" | "Divide" | "Screen" | "Overlay" | "Linear Burn" | "Difference" | "Hard Light" | "Soft Light" | "Vivid Light" | "Linear Light" | "Color Burn" | "Color Dodge"; - /** - * Opacity - * @description Desired opacity of the upper layer - * @default 1 - */ - opacity?: number; - /** - * @description Optional mask, used to restrict areas from blending - * @default null - */ - mask?: components["schemas"]["ImageField"] | null; - /** - * Fit To Width - * @description Scale upper layer to fit base width - * @default false - */ - fit_to_width?: boolean; - /** - * Fit To Height - * @description Scale upper layer to fit base height - * @default true - */ - fit_to_height?: boolean; - /** - * @description The bottom image to blend - * @default null - */ - layer_base?: components["schemas"]["ImageField"] | null; - /** - * Color Space - * @description Available color spaces for blend computations - * @default RGB - * @enum {string} - */ - color_space?: "RGB" | "Linear RGB" | "HSL (RGB)" | "HSV (RGB)" | "Okhsl" | "Okhsv" | "Oklch (Oklab)" | "LCh (CIELab)"; - /** - * Adaptive Gamut - * @description Adaptive gamut clipping (0=off). Higher prioritizes chroma over lightness - * @default 0 - */ - adaptive_gamut?: number; - /** - * High Precision - * @description Use more steps in computing gamut when possible - * @default true - */ - high_precision?: boolean; - /** - * type - * @default invokeai_img_blend - * @constant - */ - type: "invokeai_img_blend"; - }; - /** - * Image Compositor - * @description Removes backdrop from subject image then overlays subject on background image. Originally created by @dwringer - */ - InvokeImageCompositorInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Image of the subject on a plain monochrome background - * @default null - */ - image_subject?: components["schemas"]["ImageField"] | null; - /** - * @description Image of a background scene - * @default null - */ - image_background?: components["schemas"]["ImageField"] | null; - /** - * Chroma Key - * @description Can be empty for corner flood select, or CSS-3 color or tuple - * @default - */ - chroma_key?: string; - /** - * Threshold - * @description Subject isolation flood-fill threshold - * @default 50 - */ - threshold?: number; - /** - * Fill X - * @description Scale base subject image to fit background width - * @default false - */ - fill_x?: boolean; - /** - * Fill Y - * @description Scale base subject image to fit background height - * @default true - */ - fill_y?: boolean; - /** - * X Offset - * @description x-offset for the subject - * @default 0 - */ - x_offset?: number; - /** - * Y Offset - * @description y-offset for the subject - * @default 0 - */ - y_offset?: number; - /** - * type - * @default invokeai_img_composite - * @constant - */ - type: "invokeai_img_composite"; - }; - /** - * Image Dilate or Erode - * @description Dilate (expand) or erode (contract) an image. Originally created by @dwringer - */ - InvokeImageDilateOrErodeInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image from which to create a mask - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Lightness Only - * @description If true, only applies to image lightness (CIELa*b*) - * @default false - */ - lightness_only?: boolean; - /** - * Radius W - * @description Width (in pixels) by which to dilate(expand) or erode (contract) the image - * @default 4 - */ - radius_w?: number; - /** - * Radius H - * @description Height (in pixels) by which to dilate(expand) or erode (contract) the image - * @default 4 - */ - radius_h?: number; - /** - * Mode - * @description How to operate on the image - * @default Dilate - * @enum {string} - */ - mode?: "Dilate" | "Erode"; - /** - * type - * @default invokeai_img_dilate_erode - * @constant - */ - type: "invokeai_img_dilate_erode"; - }; - /** - * Enhance Image - * @description Applies processing from PIL's ImageEnhance module. Originally created by @dwringer - */ - InvokeImageEnhanceInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image for which to apply processing - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Invert - * @description Whether to invert the image colors - * @default false - */ - invert?: boolean; - /** - * Color - * @description Color enhancement factor - * @default 1 - */ - color?: number; - /** - * Contrast - * @description Contrast enhancement factor - * @default 1 - */ - contrast?: number; - /** - * Brightness - * @description Brightness enhancement factor - * @default 1 - */ - brightness?: number; - /** - * Sharpness - * @description Sharpness enhancement factor - * @default 1 - */ - sharpness?: number; - /** - * type - * @default invokeai_img_enhance - * @constant - */ - type: "invokeai_img_enhance"; - }; - /** - * Image Value Thresholds - * @description Clip image to pure black/white past specified thresholds. Originally created by @dwringer - */ - InvokeImageValueThresholdsInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image from which to create a mask - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Invert Output - * @description Make light areas dark and vice versa - * @default false - */ - invert_output?: boolean; - /** - * Renormalize Values - * @description Rescale remaining values from minimum to maximum - * @default false - */ - renormalize_values?: boolean; - /** - * Lightness Only - * @description If true, only applies to image lightness (CIELa*b*) - * @default false - */ - lightness_only?: boolean; - /** - * Threshold Upper - * @description Threshold above which will be set to full value - * @default 0.5 - */ - threshold_upper?: number; - /** - * Threshold Lower - * @description Threshold below which will be set to minimum value - * @default 0.5 - */ - threshold_lower?: number; - /** - * type - * @default invokeai_img_val_thresholds - * @constant - */ - type: "invokeai_img_val_thresholds"; - }; - /** - * ItemIdsResult - * @description Response containing ordered item ids with metadata for optimistic updates. - */ - ItemIdsResult: { - /** - * Item Ids - * @description Ordered list of item ids - */ - item_ids: number[]; - /** - * Total Count - * @description Total number of queue items matching the query - */ - total_count: number; - }; - /** - * IterateInvocation - * @description Iterates over a list of items - */ - IterateInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The list of items to iterate over - * @default [] - */ - collection?: unknown[]; - /** - * Index - * @description The index, will be provided on executed iterators - * @default 0 - */ - index?: number; - /** - * type - * @default iterate - * @constant - */ - type: "iterate"; - }; - /** - * IterateInvocationOutput - * @description Used to connect iteration outputs. Will be expanded to a specific output. - */ - IterateInvocationOutput: { - /** - * Collection Item - * @description The item being iterated over - */ - item: unknown; - /** - * Index - * @description The index of the item - */ - index: number; - /** - * Total - * @description The total number of items - */ - total: number; - /** - * type - * @default iterate_output - * @constant - */ - type: "iterate_output"; - }; - /** - * JsonListStringsOutput - * @description Output class for two strings extracted from a JSON list. - */ - JsonListStringsOutput: { - /** - * Prompt - * @description The prompt string from the JSON list - */ - prompt: string; - /** - * Seed Vector - * @description The seed vector string from the JSON list - */ - seed_vector: string; - /** - * type - * @default json_list_strings_output - * @constant - */ - type: "json_list_strings_output"; - }; - JsonValue: unknown; - /** - * LaMa Infill - * @description Infills transparent areas of an image using the LaMa model - */ - LaMaInfillInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default infill_lama - * @constant - */ - type: "infill_lama"; - }; - /** - * Latent Average - * @description Average two latents - */ - LatentAverageInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latentA?: components["schemas"]["LatentsField"] | null; - /** - * @description Latents tensor - * @default null - */ - latentB?: components["schemas"]["LatentsField"] | null; - /** - * Channels - * @description Comma-separated list of channel indices to average. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * type - * @default latent_average - * @constant - */ - type: "latent_average"; - }; - /** - * Latent Band Pass Filter - * @description Latent Band Pass Filter - */ - LatentBandPassFilterInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Input latent - * @default null - */ - input_latent?: components["schemas"]["LatentsField"] | null; - /** - * Lower Freq - * @description lower frequency (0-0.5) - * @default 0.1 - */ - lower_freq?: number; - /** - * Upper Freq - * @description upper frequency (0-0.5) - * @default 0.2 - */ - upper_freq?: number; - /** - * Normalize Std Dev - * @description normalize std dev to 1.0 - * @default false - */ - normalize_std_dev?: boolean; - /** - * type - * @default latent_band_pass_filter - * @constant - */ - type: "latent_band_pass_filter"; - }; - /** - * Latent Blend Linear - * @description Blend two latents Linearly - */ - LatentBlendLinearInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latentA?: components["schemas"]["LatentsField"] | null; - /** - * @description Latents tensor - * @default null - */ - latentB?: components["schemas"]["LatentsField"] | null; - /** - * A B Ratio - * @description Ratio of latentA to latentB (1=All A, 0=All B) - * @default 0.5 - */ - a_b_ratio?: number; - /** - * Channels - * @description Comma-separated list of channel indices to average. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * type - * @default latent_blend_linear - * @constant - */ - type: "latent_blend_linear"; - }; - /** - * Latent Channels to Grid - * @description Generates a scaled grid Images from latent channels - */ - LatentChannelsToGridInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Disable - * @description Disable this node - * @default false - */ - disable?: boolean; - /** - * @description Latents tensor - * @default null - */ - latent?: components["schemas"]["LatentsField"] | null; - /** - * Scale - * @description Overall scale factor for the grid. - * @default 1 - */ - scale?: number; - /** - * Normalize Channels - * @description Normalize all channels using a common min/max range. - * @default false - */ - normalize_channels?: boolean; - /** - * Output Bit Depth - * @description Output as 8-bit or 16-bit grayscale. - * @default 8 - * @enum {string} - */ - output_bit_depth?: "8" | "16"; - /** - * Title - * @description Title to display on the image - * @default Latent Channel Grid - */ - title?: string; - /** - * type - * @default latent_channels_to_grid - * @constant - */ - type: "latent_channels_to_grid"; - }; - /** - * Latent Combine - * @description Combines two latent tensors using various methods - */ - LatentCombineInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latentA?: components["schemas"]["LatentsField"] | null; - /** - * @description Latents tensor - * @default null - */ - latentB?: components["schemas"]["LatentsField"] | null; - /** - * Method - * @description Combination method - * @default average - * @enum {string} - */ - method?: "average" | "add" | "subtract" | "difference" | "maximum" | "minimum" | "multiply" | "frequency_blend" | "screen" | "dodge" | "burn" | "overlay" | "soft_light" | "hard_light" | "color_dodge" | "color_burn" | "linear_dodge" | "linear_burn"; - /** - * Weighted Combine - * @description Alter input latents by the provided weights - * @default false - */ - weighted_combine?: boolean; - /** - * Weight A - * @description Weight for latent A - * @default 0.5 - */ - weight_a?: number; - /** - * Weight B - * @description Weight for latent B - * @default 0.5 - */ - weight_b?: number; - /** - * Scale To Input Ranges - * @description Scale output to input ranges - * @default false - */ - scale_to_input_ranges?: boolean; - /** - * Channels - * @description Comma-separated list of channel indices to combine. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * type - * @default latent_combine - * @constant - */ - type: "latent_combine"; - }; - /** - * Latent Dtype Convert - * @description Converts the dtype of a latent tensor. - */ - LatentDtypeConvertInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latent?: components["schemas"]["LatentsField"] | null; - /** - * Dtype - * @description The target dtype for the latent tensor. - * @default bfloat16 - * @enum {string} - */ - dtype?: "float32" | "float16" | "bfloat16"; - /** - * type - * @default latent_dtype_convert - * @constant - */ - type: "latent_dtype_convert"; - }; - /** - * Latent High Pass Filter - * @description Latent High Pass Filter - */ - LatentHighPassFilterInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Input latent - * @default null - */ - input_latent?: components["schemas"]["LatentsField"] | null; - /** - * Lower Freq - * @description lower frequency (0-0.5) - * @default 0.15 - */ - lower_freq?: number; - /** - * Normalize Std Dev - * @description normalize std dev to 1.0 - * @default false - */ - normalize_std_dev?: boolean; - /** - * type - * @default latent_high_pass_filter - * @constant - */ - type: "latent_high_pass_filter"; - }; - /** - * Latent Low Pass Filter - * @description Latent Low Pass Filter - */ - LatentLowPassFilterInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Input latent - * @default null - */ - input_latent?: components["schemas"]["LatentsField"] | null; - /** - * Upper Freq - * @description upper frequency (0-0.5) - * @default 0.05 - */ - upper_freq?: number; - /** - * Normalize Std Dev - * @description normalize std dev to 1.0 - * @default false - */ - normalize_std_dev?: boolean; - /** - * type - * @default latent_low_pass_filter - * @constant - */ - type: "latent_low_pass_filter"; - }; - /** - * Latent Match - * @description Matches the distribution of one latent tensor to that of another, - * using either histogram matching or standard deviation matching. - */ - LatentMatchInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The input latent tensor to be matched. - * @default null - */ - input_latent?: components["schemas"]["LatentsField"] | null; - /** - * @description The reference latent tensor to match against. - * @default null - */ - reference_latent?: components["schemas"]["LatentsField"] | null; - /** - * Method - * @description The matching method to use. - * @default histogram - * @enum {string} - */ - method?: "histogram" | "std_dev" | "mean" | "std_dev+mean" | "cdf" | "moment" | "range" | "std_dev+mean+range"; - /** - * Channels - * @description Comma-separated list of channel indices to match. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * type - * @default latent_match - * @constant - */ - type: "latent_match"; - }; - /** - * Latent Modify - * @description Modifies selected channels of a latent tensor using scale and shift. - */ - LatentModifyChannelsInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latent?: components["schemas"]["LatentsField"] | null; - /** - * Channels - * @description Comma-separated list of channel indices to modify. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * Scale - * @description Scale factor to apply to the selected channels. - * @default 1 - */ - scale?: number; - /** - * Shift - * @description Shift value to add to the selected channels. - * @default 0 - */ - shift?: number; - /** - * type - * @default latent_modify_channels - * @constant - */ - type: "latent_modify_channels"; - }; - /** - * Latent Normalize Range - * @description Normalize a Latents Range - */ - LatentNormalizeRangeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latent?: components["schemas"]["LatentsField"] | null; - /** - * Min - * @description Min for new range - * @default -1 - */ - min?: number; - /** - * Max - * @description Max for new range - * @default 1 - */ - max?: number; - /** - * Channels - * @description Comma-separated list of channel indices to average. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * type - * @default latent_normalize_range - * @constant - */ - type: "latent_normalize_range"; - }; - /** - * Latent Normalize Std-Dev - * @description Normalize a Latents Std Dev - */ - LatentNormalizeStdDevInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latent?: components["schemas"]["LatentsField"] | null; - /** - * Std Dev - * @description Standard deviation to normalize to - * @default 1 - */ - std_dev?: number; - /** - * Channels - * @description Comma-separated list of channel indices to average. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * type - * @default latent_normalize_std_dev - * @constant - */ - type: "latent_normalize_std_dev"; - }; - /** - * Latent Normalize STD Dev and Range - * @description Normalize a Latents Range and Std Dev - */ - LatentNormalizeStdRangeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latent?: components["schemas"]["LatentsField"] | null; - /** - * Std Dev - * @description Standard deviation to normalize to - * @default 1 - */ - std_dev?: number; - /** - * Min - * @description Min for new range - * @default -1 - */ - min?: number; - /** - * Max - * @description Max for new range - * @default 1 - */ - max?: number; - /** - * Tolerance - * @description tolerance - * @default 0.001 - */ - tolerance?: number; - /** - * Max Iterations - * @description Max iterations - * @default 10 - */ - max_iterations?: number; - /** - * Channels - * @description Comma-separated list of channel indices to average. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * type - * @default latent_normalize_std_dev_range - * @constant - */ - type: "latent_normalize_std_dev_range"; - }; - /** - * Latent Plot - * @description Generates plots from latent channels and display in a grid. - */ - LatentPlotInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Disable - * @description Disable this node - * @default false - */ - disable?: boolean; - /** - * @description Latents tensor - * @default null - */ - latent?: components["schemas"]["LatentsField"] | null; - /** - * Cell X Multiplier - * @description x size multiplier for grid cells - * @default 4 - */ - cell_x_multiplier?: number; - /** - * Cell Y Multiplier - * @description y size multiplier for grid cells - * @default 3 - */ - cell_y_multiplier?: number; - /** - * Histogram Plot - * @description Plot histogram - * @default true - */ - histogram_plot?: boolean; - /** - * Histogram Bins - * @description Number of bins for the histogram - * @default 256 - */ - histogram_bins?: number; - /** - * Histogram Log Scale - * @description Use log scale for histogram y-axis - * @default false - */ - histogram_log_scale?: boolean; - /** - * Box Plot - * @description Plot box and whisker - * @default true - */ - box_plot?: boolean; - /** - * Stats Plot - * @description Plot distribution data (mean, std, mode, min, max, dtype) - * @default true - */ - stats_plot?: boolean; - /** - * Title - * @description Title to display on the image - * @default Latent Channel Plots - */ - title?: string; - /** - * Channels - * @description Comma-separated list of channel indices to plot. Use 'all' for all channels. - * @default all - */ - channels?: string; - /** - * Common Axis - * @description Use a common axis scales for all plots - * @default false - */ - common_axis?: boolean; - /** - * type - * @default latent_plot - * @constant - */ - type: "latent_plot"; - }; - /** - * Latent Quantize (Kohonen map) - * @description Use a self-organizing map to quantize the values of a latent tensor. - * - * This is highly experimental and not really suitable for most use cases. It's very easy to use settings that will appear to hang, or tie up the PC for a very long time, so use of this node is somewhat discouraged. - */ - LatentSOMInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The latents tensor to quantize - * @default null - */ - latents_in?: components["schemas"]["LatentsField"] | null; - /** - * @description Optional alternate latents to use for training - * @default null - */ - reference_in?: components["schemas"]["LatentsField"] | null; - /** - * Width - * @description Width (in cells) of the self-organizing map - * @default 4 - */ - width?: number; - /** - * Height - * @description Height (in cells) of the self-organizing map - * @default 3 - */ - height?: number; - /** - * Steps - * @description Training step count for the self-organizing map - * @default 256 - */ - steps?: number; - /** - * type - * @default latent_som - * @constant - */ - type: "latent_som"; - }; - /** - * Latent White Noise - * @description Creates White Noise latent. - */ - LatentWhiteNoiseInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Width - * @description Desired image width - * @default 512 - */ - width?: number; - /** - * Height - * @description Desired image height - * @default 512 - */ - height?: number; - /** - * Latent Type - * @description Latent Type - * @default Flux - * @enum {string} - */ - latent_type?: "Flux" | "SD3.5" | "SD/SDXL"; - /** - * Noise Type - * @description Noise Type (Uniform=random, Gausian=Normal distribution) - * @default Gausian - * @enum {string} - */ - noise_type?: "Uniform" | "Gausian"; - /** - * Seed - * @description Seed for noise generation - * @default 0 - */ - seed?: number; - /** - * type - * @default latent_white_noise - * @constant - */ - type: "latent_white_noise"; - }; - /** - * Latents Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option - */ - LatentsCollectionIndexInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Random - * @description Random Index? - * @default true - */ - random?: boolean; - /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) - * @default 0 - */ - index?: number; - /** - * Collection - * @description latents collection - * @default null - */ - collection?: components["schemas"]["LatentsField"][] | null; - /** - * type - * @default latents_collection_index - * @constant - */ - type: "latents_collection_index"; - }; - /** - * Latents Collection Primitive - * @description A collection of latents tensor primitive values - */ - LatentsCollectionInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The collection of latents tensors - * @default null - */ - collection?: components["schemas"]["LatentsField"][] | null; - /** - * type - * @default latents_collection - * @constant - */ - type: "latents_collection"; - }; - /** - * Latents Collection Primitive Linked - * @description A collection of latents tensor primitive values - */ - LatentsCollectionLinkedInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description The collection of latents tensors - * @default null - */ - collection?: components["schemas"]["LatentsField"][] | null; - /** - * type - * @default latents_collection_linked - * @constant - */ - type: "latents_collection_linked"; - /** - * @description The latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - }; - /** - * LatentsCollectionOutput - * @description Base class for nodes that output a collection of latents tensors - */ - LatentsCollectionOutput: { - /** - * Collection - * @description Latents tensor - */ - collection: components["schemas"]["LatentsField"][]; - /** - * type - * @default latents_collection_output - * @constant - */ - type: "latents_collection_output"; - }; - /** - * LatentsField - * @description A latents tensor primitive field - */ - LatentsField: { - /** - * Latents Name - * @description The name of the latents - */ - latents_name: string; - /** - * Seed - * @description Seed used to generate this latents - * @default null - */ - seed?: number | null; - }; - /** - * Latents Primitive - * @description A latents tensor primitive value - */ - LatentsInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * type - * @default latents - * @constant - */ - type: "latents"; - }; - /** - * LatentsMetaOutput - * @description Latents + metadata - */ - LatentsMetaOutput: { - /** @description Metadata Dict */ - metadata: components["schemas"]["MetadataField"]; - /** - * type - * @default latents_meta_output - * @constant - */ - type: "latents_meta_output"; - /** @description Latents tensor */ - latents: components["schemas"]["LatentsField"]; - /** - * Width - * @description Width of output (px) - */ - width: number; - /** - * Height - * @description Height of output (px) - */ - height: number; - }; - /** - * LatentsOutput - * @description Base class for nodes that output a single latents tensor - */ - LatentsOutput: { - /** @description Latents tensor */ - latents: components["schemas"]["LatentsField"]; - /** - * Width - * @description Width of output (px) - */ - width: number; - /** - * Height - * @description Height of output (px) - */ - height: number; - /** - * type - * @default latents_output - * @constant - */ - type: "latents_output"; - }; - /** - * Latents to Image - SD1.5, SDXL - * @description Generates an image from latents. - */ - LatentsToImageInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Latents tensor - * @default null - */ - latents?: components["schemas"]["LatentsField"] | null; - /** - * @description VAE - * @default null - */ - vae?: components["schemas"]["VAEField"] | null; - /** - * Tiled - * @description Processing using overlapping tiles (reduce memory consumption) - * @default false - */ - tiled?: boolean; - /** - * Tile Size - * @description The tile size for VAE tiling in pixels (image space). If set to 0, the default tile size for the model will be used. Larger tile sizes generally produce better results at the cost of higher memory usage. - * @default 0 - */ - tile_size?: number; - /** - * Fp32 - * @description Whether or not to use full float32 precision - * @default false - */ - fp32?: boolean; - /** - * type - * @default l2i - * @constant - */ - type: "l2i"; - }; - /** - * Lens Aperture Generator - * @description Generates a grayscale aperture shape with adjustable number of blades, curvature, rotation, and softness. - */ - LensApertureGeneratorInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Number Of Blades - * @description Number of aperture blades (e.g., 5 for pentagon, 6 for hexagon) - * @default 6 - */ - number_of_blades?: number; - /** - * Rounded Blades - * @description Use curved edges between aperture blades - * @default true - */ - rounded_blades?: boolean; - /** - * Curvature - * @description Aperture blade curvature (1.0 is circular) - * @default 0.5 - */ - curvature?: number; - /** - * Rotation - * @description Clockwise rotation of the aperture shape in degrees - * @default 0 - */ - rotation?: number; - /** - * Softness - * @description Softness of the aperture edge (0.0 = hard, 1.0 = feathered) - * @default 0.1 - */ - softness?: number; - /** - * type - * @default lens_aperture_generator - * @constant - */ - type: "lens_aperture_generator"; - }; - /** - * Lens Blur - * @description Adds lens blur to the input image, first converting the input image to RGB - */ - LensBlurInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to streak - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The depth map to use - * @default null - */ - depth_map?: components["schemas"]["ImageField"] | null; - /** - * @description The aperture image to use for convolution - * @default null - */ - aperture_image?: components["schemas"]["ImageField"] | null; - /** - * Focal Distance - * @description The distance at which focus is sharp: 0.0 (near) - 1.0 (far) - * @default 0.5 - */ - focal_distance?: number; - /** - * Distance Range - * @description The depth of field range around the focal distance - * @default 0.2 - */ - distance_range?: number; - /** - * Max Blur Radius - * @description Maximum blur radius - * @default 5 - */ - max_blur_radius?: number; - /** - * Max Blur Steps - * @description Number of blur maps to use internally - * @default 32 - */ - max_blur_steps?: number; - /** - * Anamorphic Factor - * @description Anamorphic squeeze factor - * @default 1.33 - */ - anamorphic_factor?: number; - /** - * Highlight Threshold - * @description The luminance threshold at which highlight enhancement begins - * @default 0.75 - */ - highlight_threshold?: number; - /** - * Highlight Factor - * @description Minimum multiplier for highlights at the threshold - * @default 1 - */ - highlight_factor?: number; - /** - * Highlight Factor High - * @description Maximum multiplier for highlights at full brightness - * @default 2 - */ - highlight_factor_high?: number; - /** - * Reduce Halos - * @description Reduce halos around areas of sharp depth distance (SLOW) - * @default true - */ - reduce_halos?: boolean; - /** - * type - * @default lens_blur - * @constant - */ - type: "lens_blur"; - }; - /** - * Lens Vignette - * @description Apply realistic lens vignetting to an image with configurable parameters. - */ - LensVignetteInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to apply vignetting to - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Intensity - * @description Intensity of the vignette effect (0-4) - * @default 0.5 - */ - intensity?: number; - /** - * Aperture Factor - * @description Aperture falloff factor (>1.2 = strong, <0.5 = weak) - * @default 0.5 - */ - aperture_factor?: number; - /** - * Center X - * @description X coordinate of vignette center (0-1) - * @default 0.5 - */ - center_x?: number; - /** - * Center Y - * @description Y coordinate of vignette center (0-1) - * @default 0.5 - */ - center_y?: number; - /** - * Preserve Highlights - * @description Preserve highlights - * @default true - */ - preserve_highlights?: boolean; - /** - * type - * @default lens_vignette - * @constant - */ - type: "lens_vignette"; - }; - /** - * Lineart Anime Edge Detection - * @description Geneartes an edge map using the Lineart model. - */ - LineartAnimeEdgeDetectionInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * type - * @default lineart_anime_edge_detection - * @constant - */ - type: "lineart_anime_edge_detection"; - }; - /** - * Lineart Edge Detection - * @description Generates an edge map using the Lineart model. - */ - LineartEdgeDetectionInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Coarse - * @description Whether to use coarse mode - * @default false - */ - coarse?: boolean; - /** - * type - * @default lineart_edge_detection - * @constant - */ - type: "lineart_edge_detection"; - }; - /** - * LLaVA OneVision VLLM - * @description Run a LLaVA OneVision VLLM model. - */ - LlavaOnevisionVllmInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Images - * @description Input image. - * @default null - */ - images?: (components["schemas"]["ImageField"][] | components["schemas"]["ImageField"]) | null; - /** - * Prompt - * @description Input text prompt. - * @default - */ - prompt?: string; - /** - * LLaVA Model Type - * @description The VLLM model to use - * @default null - */ - vllm_model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * type - * @default llava_onevision_vllm - * @constant - */ - type: "llava_onevision_vllm"; - }; - /** - * LlavaOnevision_Diffusers_Config - * @description Model config for Llava Onevision models. - */ - LlavaOnevision_Diffusers_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - /** - * Type - * @default llava_onevision - * @constant - */ - type: "llava_onevision"; - /** - * Base - * @default any - * @constant - */ - base: "any"; - /** - * Cpu Only - * @description Whether this model should run on CPU only - */ - cpu_only: boolean | null; - }; - /** - * LoRA Collection From Path - * @description Loads a collection of LoRAs filtered based on their path on disk. - */ - LoRACollectionFromPathInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRA Type - * @description The type of LoRAs that will be retrieved. - * @default sdxl - * @enum {string} - */ - model_type?: "sd-1" | "sd-2" | "sdxl" | "sdxl-refiner"; - /** - * Path Filter - * @description The path for the lora must contain the filter string - * @default null - */ - path_filter?: string | null; - /** - * Default Weight - * @description The weight to apply to each lora. This can be changed after using a Reapply LoRA Weight node. - * @default 0.75 - */ - default_weight?: number; - /** - * type - * @default lora_collection_from_path_invocation - * @constant - */ - type: "lora_collection_from_path_invocation"; - }; - /** - * LoRACollectionFromPathOutput - * @description Filtered LoRA Colleciton Output - */ - LoRACollectionFromPathOutput: { - /** - * LoRAs - * @description A collection of LoRAs. - */ - loras: components["schemas"]["LoRAField"][]; - /** - * type - * @default lora_collection_from_path_output - * @constant - */ - type: "lora_collection_from_path_output"; - }; - /** - * LoRA Collection Primitive - * @description A collection of LoRA primitive values - */ - LoRACollectionInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRAs - * @description The collection of LoRA values - * @default [] - */ - collection?: components["schemas"]["LoRAField"][]; - /** - * type - * @default lora_collection - * @constant - */ - type: "lora_collection"; - }; - /** - * LoRA Collection Primitive Linked - * @description Selects a LoRA model and weight. - */ - LoRACollectionLinkedInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRAs - * @description The collection of LoRA values - * @default [] - */ - collection?: components["schemas"]["LoRAField"][]; - /** - * type - * @default lora_collection_linked - * @constant - */ - type: "lora_collection_linked"; - /** - * LoRA - * @description LoRA model to load - * @default null - */ - lora?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 - */ - weight?: number; - }; - /** - * Apply LoRA Collection - SD1.5 - * @description Applies a collection of LoRAs to the provided UNet and CLIP models. - */ - LoRACollectionLoader: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. - * @default null - */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; - /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip?: components["schemas"]["CLIPField"] | null; - /** - * type - * @default lora_collection_loader - * @constant - */ - type: "lora_collection_loader"; - }; - /** LoRACollectionOutput */ - LoRACollectionOutput: { - /** - * LoRAs - * @description The collection of input items - */ - collection: components["schemas"]["LoRAField"][]; - /** - * type - * @default lora_collection_output - * @constant - */ - type: "lora_collection_output"; - }; - /** - * LoRA Collection Toggle - * @description Allows boolean selection between two separate LoRA collection inputs - */ - LoRACollectionToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * Col1 - * @description First LoRA Collection Input - * @default null - */ - col1?: components["schemas"]["LoRAField"][] | null; - /** - * Col2 - * @description Second LoRA Collection Input - * @default null - */ - col2?: components["schemas"]["LoRAField"][] | null; - /** - * type - * @default lora_collection_toggle - * @constant - */ - type: "lora_collection_toggle"; - }; - /** LoRACollectionToggleOutput */ - LoRACollectionToggleOutput: { - /** - * LoRA Collection - * @description Collection of LoRA models and weights - */ - collection: components["schemas"]["LoRAField"][]; - /** - * type - * @default lora_collection_toggle_output - * @constant - */ - type: "lora_collection_toggle_output"; - }; - /** LoRAField */ - LoRAField: { - /** @description Info to load lora model */ - lora: components["schemas"]["ModelIdentifierField"]; - /** - * Weight - * @description Weight to apply to lora model - */ - weight: number; - }; - /** - * Apply LoRA - SD1.5 - * @description Apply selected lora to unet and text_encoder. - */ - LoRALoaderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRA - * @description LoRA model to load - * @default null - */ - lora?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 - */ - weight?: number; - /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip?: components["schemas"]["CLIPField"] | null; - /** - * type - * @default lora_loader - * @constant - */ - type: "lora_loader"; - }; - /** - * LoRALoaderOutput - * @description Model loader output - */ - LoRALoaderOutput: { - /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null - */ - unet: components["schemas"]["UNetField"] | null; - /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip: components["schemas"]["CLIPField"] | null; - /** - * type - * @default lora_loader_output - * @constant - */ - type: "lora_loader_output"; - }; - /** - * LoRAMetadataField - * @description LoRA Metadata Field - */ - LoRAMetadataField: { - /** @description LoRA model to load */ - model: components["schemas"]["ModelIdentifierField"]; - /** - * Weight - * @description The weight at which the LoRA is applied to each model - */ - weight: number; - }; - /** - * LoRA Name Grabber - * @description Outputs the LoRA's name as a string. Use a Lora Selector node as input. - */ - LoRANameGrabberInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRA - * @description The LoRA whose name you wish to grab. Get this from a LoRA Selector node. - * @default null - */ - lora?: components["schemas"]["LoRAField"] | null; - /** - * type - * @default lora_name_grabber_invocation - * @constant - */ - type: "lora_name_grabber_invocation"; - }; - /** - * LoRANameGrabberOutput - * @description LoRA Name Grabber output - */ - LoRANameGrabberOutput: { - /** - * Name - * @description Name of the LoRA - */ - name: string; - /** - * type - * @default lora_name_grabber_output - * @constant - */ - type: "lora_name_grabber_output"; - }; - /** - * LoRARecallParameter - * @description LoRA configuration for recall - */ - LoRARecallParameter: { - /** - * Model Name - * @description The name of the LoRA model - */ - model_name: string; - /** - * Weight - * @description The weight for the LoRA - * @default 0.75 - */ - weight?: number; - /** - * Is Enabled - * @description Whether the LoRA is enabled - * @default true - */ - is_enabled?: boolean; - }; - /** - * Select LoRA - * @description Selects a LoRA model and weight. - */ - LoRASelectorInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRA - * @description LoRA model to load - * @default null - */ - lora?: components["schemas"]["ModelIdentifierField"] | null; - /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 - */ - weight?: number; - /** - * type - * @default lora_selector - * @constant - */ - type: "lora_selector"; - }; - /** - * LoRASelectorOutput - * @description Model loader output - */ - LoRASelectorOutput: { - /** - * LoRA - * @description LoRA model and weight - */ - lora: components["schemas"]["LoRAField"]; - /** - * type - * @default lora_selector_output - * @constant - */ - type: "lora_selector_output"; - }; - /** - * LoRA Toggle - * @description Allows boolean selection between two separate LoRA inputs - */ - LoRAToggleInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * @description First LoRA Input - * @default null - */ - lora1?: components["schemas"]["LoRAField"] | null; - /** - * @description Second LoRA Input - * @default null - */ - lora2?: components["schemas"]["LoRAField"] | null; - /** - * type - * @default lora_toggle - * @constant - */ - type: "lora_toggle"; - }; - /** LoRAToggleOutput */ - LoRAToggleOutput: { - /** - * LoRA - * @description LoRA model and weight - */ - lora: components["schemas"]["LoRAField"]; - /** - * type - * @default lora_toggle_output - * @constant - */ - type: "lora_toggle_output"; - }; - /** LoRA_Diffusers_FLUX_Config */ - LoRA_Diffusers_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** - * Base - * @default flux - * @constant - */ - base: "flux"; - }; - /** - * LoRA_Diffusers_Flux2_Config - * @description Model config for FLUX.2 (Klein) LoRA models in Diffusers format. - */ - LoRA_Diffusers_Flux2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** - * Base - * @default flux2 - * @constant - */ - base: "flux2"; - variant: components["schemas"]["Flux2VariantType"] | null; - }; - /** LoRA_Diffusers_SD1_Config */ - LoRA_Diffusers_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** - * Base - * @default sd-1 - * @constant - */ - base: "sd-1"; - }; - /** LoRA_Diffusers_SD2_Config */ - LoRA_Diffusers_SD2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** - * Base - * @default sd-2 - * @constant - */ - base: "sd-2"; - }; - /** LoRA_Diffusers_SDXL_Config */ - LoRA_Diffusers_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** - * Base - * @default sdxl - * @constant - */ - base: "sdxl"; - }; - /** - * LoRA_Diffusers_ZImage_Config - * @description Model config for Z-Image LoRA models in Diffusers format. - */ - LoRA_Diffusers_ZImage_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** - * Base - * @default z-image - * @constant - */ - base: "z-image"; - variant: components["schemas"]["ZImageVariantType"] | null; - }; - /** LoRA_LyCORIS_FLUX_Config */ - LoRA_LyCORIS_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default lycoris - * @constant - */ - format: "lycoris"; - /** - * Base - * @default flux - * @constant - */ - base: "flux"; - }; - /** - * LoRA_LyCORIS_Flux2_Config - * @description Model config for FLUX.2 (Klein) LoRA models in LyCORIS format. - */ - LoRA_LyCORIS_Flux2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default lycoris - * @constant - */ - format: "lycoris"; - /** - * Base - * @default flux2 - * @constant - */ - base: "flux2"; - variant: components["schemas"]["Flux2VariantType"] | null; - }; - /** LoRA_LyCORIS_SD1_Config */ - LoRA_LyCORIS_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default lycoris - * @constant - */ - format: "lycoris"; - /** - * Base - * @default sd-1 - * @constant - */ - base: "sd-1"; - }; - /** LoRA_LyCORIS_SD2_Config */ - LoRA_LyCORIS_SD2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default lycoris - * @constant - */ - format: "lycoris"; - /** - * Base - * @default sd-2 - * @constant - */ - base: "sd-2"; - }; - /** LoRA_LyCORIS_SDXL_Config */ - LoRA_LyCORIS_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default lycoris - * @constant - */ - format: "lycoris"; - /** - * Base - * @default sdxl - * @constant - */ - base: "sdxl"; - }; - /** - * LoRA_LyCORIS_ZImage_Config - * @description Model config for Z-Image LoRA models in LyCORIS format. - */ - LoRA_LyCORIS_ZImage_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default lycoris - * @constant - */ - format: "lycoris"; - /** - * Base - * @default z-image - * @constant - */ - base: "z-image"; - variant: components["schemas"]["ZImageVariantType"] | null; - }; - /** LoRA_OMI_FLUX_Config */ - LoRA_OMI_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default omi - * @constant - */ - format: "omi"; - /** - * Base - * @default flux - * @constant - */ - base: "flux"; - }; - /** LoRA_OMI_SDXL_Config */ - LoRA_OMI_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default lora - * @constant - */ - type: "lora"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; - /** - * Format - * @default omi - * @constant - */ - format: "omi"; - /** - * Base - * @default sdxl - * @constant - */ - base: "sdxl"; - }; - /** - * Load All Text Files In Folder - * @description Loads all text files in a folder and its subfolders recursively, returning them as a Collection of strings - */ - LoadAllTextFilesInFolderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * File Extension - * @description Only files with the given extension will be loaded. For example: json - * @default null - */ - extension_to_match?: string | null; - /** - * Folder Path - * @description The path to load files from. - * @default null - */ - folder_path?: string | null; - /** - * type - * @default load_all_text_files_in_folder_output - * @constant - */ - type: "load_all_text_files_in_folder_output"; - }; - /** - * LoadAllTextFilesInFolderOutput - * @description Load Text Files In Folder Output - */ - LoadAllTextFilesInFolderOutput: { - /** Results */ - result: string[]; - /** - * type - * @default load_all_text_files_in_folder_output - * @constant - */ - type: "load_all_text_files_in_folder_output"; - }; - /** - * Load Aperture Image - * @description Loads a grayscale aperture image from the presets directory. - */ - LoadApertureImageInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Aperture Image - * @description The aperture image to load from the 'aperture_images' folder - * @default 5_point_star.png - * @enum {string} - */ - aperture_image?: "5_point_star.png" | "cross.png" | "diag_arrow.png" | "diffraction.png" | "heart.png"; - /** - * type - * @default load_aperture_image - * @constant - */ - type: "load_aperture_image"; - }; - /** - * Load Text File to String - * @description Loads a text from a provided path and outputs it as a string - */ - LoadTextFileToStringInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Path - * @description The full path to the text file. - * @default null - */ - file_path?: string | null; - /** - * type - * @default load_text_file_to_string_invocation - * @constant - */ - type: "load_text_file_to_string_invocation"; - }; - /** - * LoadTextFileToStringOutput - * @description Load Text File To String Output - */ - LoadTextFileToStringOutput: { - /** Result */ - result: string; - /** - * type - * @default load_text_file_to_string_output - * @constant - */ - type: "load_text_file_to_string_output"; - }; - /** - * Load Video Frame - * @description Load a specific frame from an MP4 video and provide it as output. - */ - LoadVideoFrameInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Video Path - * @description The path to the MP4 video file - * @default null - */ - video_path?: string | null; - /** - * Frame Number - * @description The frame number to load - * @default 1 - */ - frame_number?: number; - /** - * type - * @default load_video_frame - * @constant - */ - type: "load_video_frame"; - }; - /** - * LocalModelSource - * @description A local file or directory path. - */ - LocalModelSource: { - /** Path */ - path: string; - /** - * Inplace - * @default false - */ - inplace?: boolean | null; - /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} - */ - type: "local"; - }; - /** - * LogLevel - * @enum {integer} - */ - LogLevel: 0 | 10 | 20 | 30 | 40 | 50; - /** - * LoginRequest - * @description Request body for user login. - */ - LoginRequest: { - /** - * Email - * @description User email address - */ - email: string; - /** - * Password - * @description User password - */ - password: string; - /** - * Remember Me - * @description Whether to extend session duration - * @default false - */ - remember_me?: boolean; - }; - /** - * LoginResponse - * @description Response from successful login. - */ - LoginResponse: { - /** - * Token - * @description JWT access token - */ - token: string; - /** @description User information */ - user: components["schemas"]["UserDTO"]; - /** - * Expires In - * @description Token expiration time in seconds - */ - expires_in: number; - }; - /** - * LogoutResponse - * @description Response from logout. - */ - LogoutResponse: { - /** - * Success - * @description Whether logout was successful - */ - success: boolean; - }; - /** - * Lookup LoRA Collection Triggers - * @description Retrieves trigger words from the Model Manager for all LoRAs in the collection - */ - LookupLoRACollectionTriggersInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRA Collection - * @description The LoRAs to look up - * @default null - */ - lora_collection?: components["schemas"]["LoRAField"][] | null; - /** - * type - * @default lookup_lora_collection_triggers_invocation - * @constant - */ - type: "lookup_lora_collection_triggers_invocation"; - }; - /** - * LookupLoRACollectionTriggersOutput - * @description Lookup LoRA Collection Triggers Output - */ - LookupLoRACollectionTriggersOutput: { - /** - * Trigger Words - * @description A collection of all the LoRAs trigger words - */ - trigger_words: string[]; - /** - * type - * @default lookup_lora_collection_triggers_output - * @constant - */ - type: "lookup_lora_collection_triggers_output"; - }; - /** - * Lookup LoRA Triggers - * @description Retrieves a LoRA's trigger words from the Model Manager - */ - LookupLoRATriggersInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * LoRA - * @description The LoRA to look up - * @default null - */ - lora?: components["schemas"]["LoRAField"] | null; - /** - * type - * @default lookup_lora_triggers_invocation - * @constant - */ - type: "lookup_lora_triggers_invocation"; - }; - /** - * LookupLoRATriggersOutput - * @description Lookup LoRA Triggers Output - */ - LookupLoRATriggersOutput: { - /** - * Trigger Words - * @description A collection of the LoRAs trigger words - */ - trigger_words: string[]; - /** - * type - * @default lookup_lora_triggers_output - * @constant - */ - type: "lookup_lora_triggers_output"; - }; - /** - * Lookup Table from File - * @description Loads a lookup table from a YAML file - */ - LookupTableFromFileInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * File Path - * @description Path to lookup table YAML file - * @default null - */ - file_path?: string | null; - /** - * type - * @default lookup_table_from_file - * @constant - */ - type: "lookup_table_from_file"; - }; - /** - * LookupTableOutput - * @description Base class for invocations that output a JSON lookup table - */ - LookupTableOutput: { - /** - * Lookups - * @description The output lookup table - */ - lookups: string; - /** - * type - * @default lookups_output - * @constant - */ - type: "lookups_output"; - }; - /** - * Lookups Entry from Prompt - * @description Creates a lookup table of a single heading->value - */ - LookupsEntryFromPromptInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Heading - * @description Heading for the lookup table entry - * @default null - */ - heading?: string | null; - /** - * Lookup - * @description The entry to place under Heading in the lookup table - * @default - */ - lookup?: string; - /** - * type - * @default lookup_from_prompt - * @constant - */ - type: "lookup_from_prompt"; - }; - /** LoraModelDefaultSettings */ - LoraModelDefaultSettings: { - /** - * Weight - * @description Default weight for this model - */ - weight?: number | null; - }; - /** - * LoRA To String - * @description Converts an lora Model to a JSONString - */ - LoraToStringInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description LoRA model to load - * @default null - */ - model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * type - * @default lora_to_string - * @constant - */ - type: "lora_to_string"; - }; - /** MDControlListOutput */ - MDControlListOutput: { - /** - * ControlNet-List - * @description ControlNet(s) to apply - */ - control_list: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; - /** - * type - * @default md_control_list_output - * @constant - */ - type: "md_control_list_output"; - }; - /** MDIPAdapterListOutput */ - MDIPAdapterListOutput: { - /** - * IP-Adapter-List - * @description IP-Adapter to apply - */ - ip_adapter_list: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; - /** - * type - * @default md_ip_adapter_list_output - * @constant - */ - type: "md_ip_adapter_list_output"; - }; - /** MDT2IAdapterListOutput */ - MDT2IAdapterListOutput: { - /** - * T2I Adapter-List - * @description T2I-Adapter(s) to apply - */ - t2i_adapter_list: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; - /** - * type - * @default md_ip_adapters_output - * @constant - */ - type: "md_ip_adapters_output"; - }; - /** - * MLSD Detection - * @description Generates an line segment map using MLSD. - */ - MLSDDetectionInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Score Threshold - * @description The threshold used to score points when determining line segments - * @default 0.1 - */ - score_threshold?: number; - /** - * Distance Threshold - * @description Threshold for including a line segment - lines shorter than this distance will be discarded - * @default 20 - */ - distance_threshold?: number; - /** - * type - * @default mlsd_detection - * @constant - */ - type: "mlsd_detection"; - }; - /** MainModelDefaultSettings */ - MainModelDefaultSettings: { - /** - * Vae - * @description Default VAE for this model (model key) - */ - vae?: string | null; - /** - * Vae Precision - * @description Default VAE precision for this model - */ - vae_precision?: ("fp16" | "fp32") | null; - /** - * Scheduler - * @description Default scheduler for this model - */ - scheduler?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; - /** - * Steps - * @description Default number of steps for this model - */ - steps?: number | null; - /** - * Cfg Scale - * @description Default CFG Scale for this model - */ - cfg_scale?: number | null; - /** - * Cfg Rescale Multiplier - * @description Default CFG Rescale Multiplier for this model - */ - cfg_rescale_multiplier?: number | null; - /** - * Width - * @description Default width for this model - */ - width?: number | null; - /** - * Height - * @description Default height for this model - */ - height?: number | null; - /** - * Guidance - * @description Default Guidance for this model - */ - guidance?: number | null; - /** - * Cpu Only - * @description Whether this model should run on CPU only - */ - cpu_only?: boolean | null; - }; - /** - * Main Model Input - * @description Loads a main model from an input, outputting its submodels. - */ - MainModelLoaderInputInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Main model (UNet, VAE, CLIP) to load - * @default null - */ - model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * type - * @default main_model_loader_input - * @constant - */ - type: "main_model_loader_input"; - }; - /** - * Main Model - SD1.5, SD2 - * @description Loads a main model, outputting its submodels. - */ - MainModelLoaderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Main model (UNet, VAE, CLIP) to load - * @default null - */ - model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * type - * @default main_model_loader - * @constant - */ - type: "main_model_loader"; - }; - /** - * Main Model To String - * @description Converts a Main Model to a JSONString - */ - MainModelToStringInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Main model (UNet, VAE, CLIP) to load - * @default null - */ - model?: components["schemas"]["ModelIdentifierField"] | null; - /** - * type - * @default main_model_to_string - * @constant - */ - type: "main_model_to_string"; - }; - /** - * Main_BnBNF4_FLUX_Config - * @description Model config for main checkpoint models. - */ - Main_BnBNF4_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Base - * @default flux - * @constant - */ - base: "flux"; - /** - * Format - * @default bnb_quantized_nf4b - * @constant - */ - format: "bnb_quantized_nf4b"; - variant: components["schemas"]["FluxVariantType"]; - }; - /** - * Main_Checkpoint_FLUX_Config - * @description Model config for main checkpoint models. - */ - Main_Checkpoint_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - /** - * Base - * @default flux - * @constant - */ - base: "flux"; - variant: components["schemas"]["FluxVariantType"]; - }; - /** - * Main_Checkpoint_Flux2_Config - * @description Model config for FLUX.2 checkpoint models (e.g. Klein). - */ - Main_Checkpoint_Flux2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - /** - * Base - * @default flux2 - * @constant - */ - base: "flux2"; - variant: components["schemas"]["Flux2VariantType"]; - }; - /** Main_Checkpoint_SD1_Config */ - Main_Checkpoint_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; - /** - * Base - * @default sd-1 - * @constant - */ - base: "sd-1"; - }; - /** Main_Checkpoint_SD2_Config */ - Main_Checkpoint_SD2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; - /** - * Base - * @default sd-2 - * @constant - */ - base: "sd-2"; - }; - /** Main_Checkpoint_SDXLRefiner_Config */ - Main_Checkpoint_SDXLRefiner_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; - /** - * Base - * @default sdxl-refiner - * @constant - */ - base: "sdxl-refiner"; - }; - /** Main_Checkpoint_SDXL_Config */ - Main_Checkpoint_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; - /** - * Base - * @default sdxl - * @constant - */ - base: "sdxl"; - }; - /** - * Main_Checkpoint_ZImage_Config - * @description Model config for Z-Image single-file checkpoint models (safetensors, etc). - */ - Main_Checkpoint_ZImage_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Base - * @default z-image - * @constant - */ - base: "z-image"; - /** - * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - variant: components["schemas"]["ZImageVariantType"]; - }; - /** Main_Diffusers_CogView4_Config */ - Main_Diffusers_CogView4_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - /** - * Base - * @default cogview4 - * @constant - */ - base: "cogview4"; - }; - /** - * Main_Diffusers_FLUX_Config - * @description Model config for FLUX.1 models in diffusers format. - */ - Main_Diffusers_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - /** - * Base - * @default flux - * @constant - */ - base: "flux"; - variant: components["schemas"]["FluxVariantType"]; - }; - /** - * Main_Diffusers_Flux2_Config - * @description Model config for FLUX.2 models in diffusers format (e.g. FLUX.2 Klein). - */ - Main_Diffusers_Flux2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - /** - * Base - * @default flux2 - * @constant - */ - base: "flux2"; - variant: components["schemas"]["Flux2VariantType"]; - }; - /** Main_Diffusers_SD1_Config */ - Main_Diffusers_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; - /** - * Base - * @default sd-1 - * @constant - */ - base: "sd-1"; - }; - /** Main_Diffusers_SD2_Config */ - Main_Diffusers_SD2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; - /** - * Base - * @default sd-2 - * @constant - */ - base: "sd-2"; - }; - /** Main_Diffusers_SD3_Config */ - Main_Diffusers_SD3_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - /** - * Base - * @default sd-3 - * @constant - */ - base: "sd-3"; - /** - * Submodels - * @description Loadable submodels in this model - */ - submodels: { - [key: string]: components["schemas"]["SubmodelDefinition"]; - } | null; - }; - /** Main_Diffusers_SDXLRefiner_Config */ - Main_Diffusers_SDXLRefiner_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; - /** - * Base - * @default sdxl-refiner - * @constant - */ - base: "sdxl-refiner"; - }; - /** Main_Diffusers_SDXL_Config */ - Main_Diffusers_SDXL_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - prediction_type: components["schemas"]["SchedulerPredictionType"]; - variant: components["schemas"]["ModelVariantType"]; - /** - * Base - * @default sdxl - * @constant - */ - base: "sdxl"; - }; - /** - * Main_Diffusers_ZImage_Config - * @description Model config for Z-Image diffusers models (Z-Image-Turbo, Z-Image-Base). - */ - Main_Diffusers_ZImage_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - /** - * Base - * @default z-image - * @constant - */ - base: "z-image"; - variant: components["schemas"]["ZImageVariantType"]; - }; - /** - * Main_GGUF_FLUX_Config - * @description Model config for main checkpoint models. - */ - Main_GGUF_FLUX_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Base - * @default flux - * @constant - */ - base: "flux"; - /** - * Format - * @default gguf_quantized - * @constant - */ - format: "gguf_quantized"; - variant: components["schemas"]["FluxVariantType"]; - }; - /** - * Main_GGUF_Flux2_Config - * @description Model config for GGUF-quantized FLUX.2 checkpoint models (e.g. Klein). - */ - Main_GGUF_Flux2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Base - * @default flux2 - * @constant - */ - base: "flux2"; - /** - * Format - * @default gguf_quantized - * @constant - */ - format: "gguf_quantized"; - variant: components["schemas"]["Flux2VariantType"]; - }; - /** - * Main_GGUF_ZImage_Config - * @description Model config for GGUF-quantized Z-Image transformer models. - */ - Main_GGUF_ZImage_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Type - * @default main - * @constant - */ - type: "main"; - /** - * Trigger Phrases - * @description Set of trigger phrases for this model - */ - trigger_phrases: string[] | null; - /** @description Default settings for this model */ - default_settings: components["schemas"]["MainModelDefaultSettings"] | null; - /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Base - * @default z-image - * @constant - */ - base: "z-image"; - /** - * Format - * @default gguf_quantized - * @constant - */ - format: "gguf_quantized"; - variant: components["schemas"]["ZImageVariantType"]; - }; - /** - * Combine Masks - * @description Combine two masks together by multiplying them using `PIL.ImageChops.multiply()`. - */ - MaskCombineInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The first mask to combine - * @default null - */ - mask1?: components["schemas"]["ImageField"] | null; - /** - * @description The second image to combine - * @default null - */ - mask2?: components["schemas"]["ImageField"] | null; - /** - * type - * @default mask_combine - * @constant - */ - type: "mask_combine"; - }; - /** - * Mask Edge - * @description Applies an edge mask to an image - */ - MaskEdgeInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to apply the mask to - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Edge Size - * @description The size of the edge - * @default null - */ - edge_size?: number | null; - /** - * Edge Blur - * @description The amount of blur on the edge - * @default null - */ - edge_blur?: number | null; - /** - * Low Threshold - * @description First threshold for the hysteresis procedure in Canny edge detection - * @default null - */ - low_threshold?: number | null; - /** - * High Threshold - * @description Second threshold for the hysteresis procedure in Canny edge detection - * @default null - */ - high_threshold?: number | null; - /** - * type - * @default mask_edge - * @constant - */ - type: "mask_edge"; - }; - /** - * Mask from Alpha - * @description Extracts the alpha channel of an image as a mask. - */ - MaskFromAlphaInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to create the mask from - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Invert - * @description Whether or not to invert the mask - * @default false - */ - invert?: boolean; - /** - * type - * @default tomask - * @constant - */ - type: "tomask"; - }; - /** - * Mask from Segmented Image - * @description Generate a mask for a particular color in an ID Map - */ - MaskFromIDInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to create the mask from - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description ID color to mask - * @default null - */ - color?: components["schemas"]["ColorField"] | null; - /** - * Threshold - * @description Threshold for color detection - * @default 100 - */ - threshold?: number; - /** - * Invert - * @description Whether or not to invert the mask - * @default false - */ - invert?: boolean; - /** - * type - * @default mask_from_id - * @constant - */ - type: "mask_from_id"; - }; - /** - * MaskOutput - * @description A torch mask tensor. - */ - MaskOutput: { - /** @description The mask. */ - mask: components["schemas"]["TensorField"]; - /** - * Width - * @description The width of the mask in pixels. - */ - width: number; - /** - * Height - * @description The height of the mask in pixels. - */ - height: number; - /** - * type - * @default mask_output - * @constant - */ - type: "mask_output"; - }; - /** - * Tensor Mask to Image - * @description Convert a mask tensor to an image. - */ - MaskTensorToImageInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The mask tensor to convert. - * @default null - */ - mask?: components["schemas"]["TensorField"] | null; - /** - * type - * @default tensor_mask_to_image - * @constant - */ - type: "tensor_mask_to_image"; - }; - /** - * Match Histogram (YCbCr) - * @description Match a histogram from one image to another using YCbCr color space - */ - MatchHistogramInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to receive the histogram - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The reference image with the source histogram - * @default null - */ - reference_image?: components["schemas"]["ImageField"] | null; - /** - * Match Luminance Only - * @description Only transfer the luminance - * @default false - */ - match_luminance_only?: boolean; - /** - * Output Grayscale - * @description Convert output image to grayscale - * @default false - */ - output_grayscale?: boolean; - /** - * type - * @default match_histogram - * @constant - */ - type: "match_histogram"; - }; - /** - * Match Histogram LAB - * @description Match a histogram from one image to another using Lab color space - */ - MatchHistogramLabInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to receive the histogram - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * @description The reference image with the source histogram - * @default null - */ - reference_image?: components["schemas"]["ImageField"] | null; - /** - * Match Luminance Only - * @description Only transfer the luminance/brightness channel - * @default false - */ - match_luminance_only?: boolean; - /** - * Output Grayscale - * @description Convert output image to grayscale - * @default false - */ - output_grayscale?: boolean; - /** - * type - * @default match_histogram_lab - * @constant - */ - type: "match_histogram_lab"; - }; - /** - * MathEval - * @description Performs arbitrary user-defined calculations on x, y, z, and w variables - */ - MathEvalInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * X - * @description Input x - * @default 0 - */ - x?: number; - /** - * Y - * @description Input y - * @default 0 - */ - y?: number; - /** - * Z - * @description Input z - * @default 0 - */ - z?: number; - /** - * W - * @description Input w - * @default 0 - */ - w?: number; - /** - * Equation A - * @description A basic mathematical equation for a involving x, y, z, and w - * @default math.sin(x) - */ - equation_a?: string; - /** - * Equation B - * @description A basic mathematical equation for b involving x, y, z, and w - * @default 1+(x*y) - */ - equation_b?: string; - /** - * Equation C - * @description A basic mathematical equation for c involving x, y, z, and w - * @default - */ - equation_c?: string; - /** - * Equation D - * @description A basic mathematical equation for d involving x, y, z, and w - * @default - */ - equation_d?: string; - /** - * type - * @default math_eval - * @constant - */ - type: "math_eval"; - }; - /** - * MathEvalOutput - * @description Base class for math output - */ - MathEvalOutput: { - /** - * A - * @description Output a - */ - a: number; - /** - * B - * @description Output b - */ - b: number; - /** - * C - * @description Output c - */ - c: number; - /** - * D - * @description Output d - */ - d: number; - /** - * type - * @default math_eval_output - * @constant - */ - type: "math_eval_output"; - }; - /** - * MediaPipe Face Detection - * @description Detects faces using MediaPipe. - */ - MediaPipeFaceDetectionInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description The image to process - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Max Faces - * @description Maximum number of faces to detect - * @default 1 - */ - max_faces?: number; - /** - * Min Confidence - * @description Minimum confidence for face detection - * @default 0.5 - */ - min_confidence?: number; - /** - * type - * @default mediapipe_face_detection - * @constant - */ - type: "mediapipe_face_detection"; - }; - /** - * Merge LoRA Collections - * @description Merges two Collections of LoRAs into a single Collection. - */ - MergeLoRACollectionsInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection 1 - * @description A collection of LoRAs or a single LoRA from a LoRA Selector node. - * @default null - */ - collection1?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; - /** - * Collection 2 - * @description A collection of LoRAs or a single LoRA from a LoRA Selector node. - * @default null - */ - collection2?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; - /** - * type - * @default merge_lora_collections_invocation - * @constant - */ - type: "merge_lora_collections_invocation"; - }; - /** - * MergeLoRACollectionsOutput - * @description Join LoRA Collections output - */ - MergeLoRACollectionsOutput: { - /** - * Collection - * @description The merged collection - */ - collection: components["schemas"]["LoRAField"][]; - /** - * type - * @default merge_lora_collections_output - * @constant - */ - type: "merge_lora_collections_output"; - }; - /** - * Metadata Merge - * @description Merged a collection of MetadataDict into a single MetadataDict. - */ - MergeMetadataInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * Collection - * @description Collection of Metadata - * @default null - */ - collection?: components["schemas"]["MetadataField"][] | null; - /** - * type - * @default merge_metadata - * @constant - */ - type: "merge_metadata"; + error_traceback: string; + }; + InvocationOutputMap: { + add: components["schemas"]["IntegerOutput"]; + alpha_mask_to_tensor: components["schemas"]["MaskOutput"]; + apply_mask_to_image: components["schemas"]["ImageOutput"]; + apply_tensor_mask_to_image: components["schemas"]["ImageOutput"]; + blank_image: components["schemas"]["ImageOutput"]; + boolean: components["schemas"]["BooleanOutput"]; + boolean_collection: components["schemas"]["BooleanCollectionOutput"]; + bounding_box: components["schemas"]["BoundingBoxOutput"]; + calculate_image_tiles: components["schemas"]["CalculateImageTilesOutput"]; + calculate_image_tiles_even_split: components["schemas"]["CalculateImageTilesOutput"]; + calculate_image_tiles_min_overlap: components["schemas"]["CalculateImageTilesOutput"]; + canny_edge_detection: components["schemas"]["ImageOutput"]; + canvas_paste_back: components["schemas"]["ImageOutput"]; + canvas_v2_mask_and_crop: components["schemas"]["ImageOutput"]; + clip_skip: components["schemas"]["CLIPSkipInvocationOutput"]; + cogview4_denoise: components["schemas"]["LatentsOutput"]; + cogview4_i2l: components["schemas"]["LatentsOutput"]; + cogview4_l2i: components["schemas"]["ImageOutput"]; + cogview4_model_loader: components["schemas"]["CogView4ModelLoaderOutput"]; + cogview4_text_encoder: components["schemas"]["CogView4ConditioningOutput"]; + collect: components["schemas"]["CollectInvocationOutput"]; + color: components["schemas"]["ColorOutput"]; + color_correct: components["schemas"]["ImageOutput"]; + color_map: components["schemas"]["ImageOutput"]; + compel: components["schemas"]["ConditioningOutput"]; + conditioning: components["schemas"]["ConditioningOutput"]; + conditioning_collection: components["schemas"]["ConditioningCollectionOutput"]; + content_shuffle: components["schemas"]["ImageOutput"]; + controlnet: components["schemas"]["ControlOutput"]; + core_metadata: components["schemas"]["MetadataOutput"]; + create_denoise_mask: components["schemas"]["DenoiseMaskOutput"]; + create_gradient_mask: components["schemas"]["GradientMaskOutput"]; + crop_image_to_bounding_box: components["schemas"]["ImageOutput"]; + crop_latents: components["schemas"]["LatentsOutput"]; + cv_inpaint: components["schemas"]["ImageOutput"]; + decode_watermark: components["schemas"]["StringOutput"]; + denoise_latents: components["schemas"]["LatentsOutput"]; + denoise_latents_meta: components["schemas"]["LatentsMetaOutput"]; + depth_anything_depth_estimation: components["schemas"]["ImageOutput"]; + div: components["schemas"]["IntegerOutput"]; + dw_openpose_detection: components["schemas"]["ImageOutput"]; + dynamic_prompt: components["schemas"]["StringCollectionOutput"]; + esrgan: components["schemas"]["ImageOutput"]; + expand_mask_with_fade: components["schemas"]["ImageOutput"]; + face_identifier: components["schemas"]["ImageOutput"]; + face_mask_detection: components["schemas"]["FaceMaskOutput"]; + face_off: components["schemas"]["FaceOffOutput"]; + float: components["schemas"]["FloatOutput"]; + float_batch: components["schemas"]["FloatOutput"]; + float_collection: components["schemas"]["FloatCollectionOutput"]; + float_generator: components["schemas"]["FloatGeneratorOutput"]; + float_math: components["schemas"]["FloatOutput"]; + float_range: components["schemas"]["FloatCollectionOutput"]; + float_to_int: components["schemas"]["IntegerOutput"]; + flux2_denoise: components["schemas"]["LatentsOutput"]; + flux2_klein_lora_collection_loader: components["schemas"]["Flux2KleinLoRALoaderOutput"]; + flux2_klein_lora_loader: components["schemas"]["Flux2KleinLoRALoaderOutput"]; + flux2_klein_model_loader: components["schemas"]["Flux2KleinModelLoaderOutput"]; + flux2_klein_text_encoder: components["schemas"]["FluxConditioningOutput"]; + flux2_vae_decode: components["schemas"]["ImageOutput"]; + flux2_vae_encode: components["schemas"]["LatentsOutput"]; + flux_control_lora_loader: components["schemas"]["FluxControlLoRALoaderOutput"]; + flux_controlnet: components["schemas"]["FluxControlNetOutput"]; + flux_denoise: components["schemas"]["LatentsOutput"]; + flux_denoise_meta: components["schemas"]["LatentsMetaOutput"]; + flux_fill: components["schemas"]["FluxFillOutput"]; + flux_ip_adapter: components["schemas"]["IPAdapterOutput"]; + flux_kontext: components["schemas"]["FluxKontextOutput"]; + flux_kontext_image_prep: components["schemas"]["ImageOutput"]; + flux_lora_collection_loader: components["schemas"]["FluxLoRALoaderOutput"]; + flux_lora_loader: components["schemas"]["FluxLoRALoaderOutput"]; + flux_model_loader: components["schemas"]["FluxModelLoaderOutput"]; + flux_redux: components["schemas"]["FluxReduxOutput"]; + flux_text_encoder: components["schemas"]["FluxConditioningOutput"]; + flux_vae_decode: components["schemas"]["ImageOutput"]; + flux_vae_encode: components["schemas"]["LatentsOutput"]; + freeu: components["schemas"]["UNetOutput"]; + get_image_mask_bounding_box: components["schemas"]["BoundingBoxOutput"]; + grounding_dino: components["schemas"]["BoundingBoxCollectionOutput"]; + hed_edge_detection: components["schemas"]["ImageOutput"]; + heuristic_resize: components["schemas"]["ImageOutput"]; + i2l: components["schemas"]["LatentsOutput"]; + ideal_size: components["schemas"]["IdealSizeOutput"]; + image: components["schemas"]["ImageOutput"]; + image_batch: components["schemas"]["ImageOutput"]; + image_collection: components["schemas"]["ImageCollectionOutput"]; + image_generator: components["schemas"]["ImageGeneratorOutput"]; + image_mask_to_tensor: components["schemas"]["MaskOutput"]; + image_panel_layout: components["schemas"]["ImagePanelCoordinateOutput"]; + img_blur: components["schemas"]["ImageOutput"]; + img_chan: components["schemas"]["ImageOutput"]; + img_channel_multiply: components["schemas"]["ImageOutput"]; + img_channel_offset: components["schemas"]["ImageOutput"]; + img_conv: components["schemas"]["ImageOutput"]; + img_crop: components["schemas"]["ImageOutput"]; + img_hue_adjust: components["schemas"]["ImageOutput"]; + img_ilerp: components["schemas"]["ImageOutput"]; + img_lerp: components["schemas"]["ImageOutput"]; + img_mul: components["schemas"]["ImageOutput"]; + img_noise: components["schemas"]["ImageOutput"]; + img_nsfw: components["schemas"]["ImageOutput"]; + img_pad_crop: components["schemas"]["ImageOutput"]; + img_paste: components["schemas"]["ImageOutput"]; + img_resize: components["schemas"]["ImageOutput"]; + img_scale: components["schemas"]["ImageOutput"]; + img_watermark: components["schemas"]["ImageOutput"]; + infill_cv2: components["schemas"]["ImageOutput"]; + infill_lama: components["schemas"]["ImageOutput"]; + infill_patchmatch: components["schemas"]["ImageOutput"]; + infill_rgba: components["schemas"]["ImageOutput"]; + infill_tile: components["schemas"]["ImageOutput"]; + integer: components["schemas"]["IntegerOutput"]; + integer_batch: components["schemas"]["IntegerOutput"]; + integer_collection: components["schemas"]["IntegerCollectionOutput"]; + integer_generator: components["schemas"]["IntegerGeneratorOutput"]; + integer_math: components["schemas"]["IntegerOutput"]; + invert_tensor_mask: components["schemas"]["MaskOutput"]; + invokeai_ealightness: components["schemas"]["ImageOutput"]; + invokeai_img_blend: components["schemas"]["ImageOutput"]; + invokeai_img_composite: components["schemas"]["ImageOutput"]; + invokeai_img_dilate_erode: components["schemas"]["ImageOutput"]; + invokeai_img_enhance: components["schemas"]["ImageOutput"]; + invokeai_img_hue_adjust_plus: components["schemas"]["ImageOutput"]; + invokeai_img_val_thresholds: components["schemas"]["ImageOutput"]; + ip_adapter: components["schemas"]["IPAdapterOutput"]; + iterate: components["schemas"]["IterateInvocationOutput"]; + l2i: components["schemas"]["ImageOutput"]; + latents: components["schemas"]["LatentsOutput"]; + latents_collection: components["schemas"]["LatentsCollectionOutput"]; + lblend: components["schemas"]["LatentsOutput"]; + lineart_anime_edge_detection: components["schemas"]["ImageOutput"]; + lineart_edge_detection: components["schemas"]["ImageOutput"]; + llava_onevision_vllm: components["schemas"]["StringOutput"]; + lora_collection_loader: components["schemas"]["LoRALoaderOutput"]; + lora_loader: components["schemas"]["LoRALoaderOutput"]; + lora_selector: components["schemas"]["LoRASelectorOutput"]; + lresize: components["schemas"]["LatentsOutput"]; + lscale: components["schemas"]["LatentsOutput"]; + main_model_loader: components["schemas"]["ModelLoaderOutput"]; + mask_combine: components["schemas"]["ImageOutput"]; + mask_edge: components["schemas"]["ImageOutput"]; + mask_from_id: components["schemas"]["ImageOutput"]; + mediapipe_face_detection: components["schemas"]["ImageOutput"]; + merge_metadata: components["schemas"]["MetadataOutput"]; + merge_tiles_to_image: components["schemas"]["ImageOutput"]; + metadata: components["schemas"]["MetadataOutput"]; + metadata_field_extractor: components["schemas"]["StringOutput"]; + metadata_from_image: components["schemas"]["MetadataOutput"]; + metadata_item: components["schemas"]["MetadataItemOutput"]; + metadata_item_linked: components["schemas"]["MetadataOutput"]; + metadata_to_bool: components["schemas"]["BooleanOutput"]; + metadata_to_bool_collection: components["schemas"]["BooleanCollectionOutput"]; + metadata_to_controlnets: components["schemas"]["MDControlListOutput"]; + metadata_to_float: components["schemas"]["FloatOutput"]; + metadata_to_float_collection: components["schemas"]["FloatCollectionOutput"]; + metadata_to_integer: components["schemas"]["IntegerOutput"]; + metadata_to_integer_collection: components["schemas"]["IntegerCollectionOutput"]; + metadata_to_ip_adapters: components["schemas"]["MDIPAdapterListOutput"]; + metadata_to_lora_collection: components["schemas"]["MetadataToLorasCollectionOutput"]; + metadata_to_loras: components["schemas"]["LoRALoaderOutput"]; + metadata_to_model: components["schemas"]["MetadataToModelOutput"]; + metadata_to_scheduler: components["schemas"]["SchedulerOutput"]; + metadata_to_sdlx_loras: components["schemas"]["SDXLLoRALoaderOutput"]; + metadata_to_sdxl_model: components["schemas"]["MetadataToSDXLModelOutput"]; + metadata_to_string: components["schemas"]["StringOutput"]; + metadata_to_string_collection: components["schemas"]["StringCollectionOutput"]; + metadata_to_t2i_adapters: components["schemas"]["MDT2IAdapterListOutput"]; + metadata_to_vae: components["schemas"]["VAEOutput"]; + mlsd_detection: components["schemas"]["ImageOutput"]; + model_identifier: components["schemas"]["ModelIdentifierOutput"]; + mul: components["schemas"]["IntegerOutput"]; + noise: components["schemas"]["NoiseOutput"]; + normal_map: components["schemas"]["ImageOutput"]; + pair_tile_image: components["schemas"]["PairTileImageOutput"]; + paste_image_into_bounding_box: components["schemas"]["ImageOutput"]; + pbr_maps: components["schemas"]["PBRMapsOutput"]; + pidi_edge_detection: components["schemas"]["ImageOutput"]; + prompt_from_file: components["schemas"]["StringCollectionOutput"]; + prompt_template: components["schemas"]["PromptTemplateOutput"]; + rand_float: components["schemas"]["FloatOutput"]; + rand_int: components["schemas"]["IntegerOutput"]; + random_range: components["schemas"]["IntegerCollectionOutput"]; + range: components["schemas"]["IntegerCollectionOutput"]; + range_of_size: components["schemas"]["IntegerCollectionOutput"]; + rectangle_mask: components["schemas"]["MaskOutput"]; + round_float: components["schemas"]["FloatOutput"]; + save_image: components["schemas"]["ImageOutput"]; + scheduler: components["schemas"]["SchedulerOutput"]; + sd3_denoise: components["schemas"]["LatentsOutput"]; + sd3_i2l: components["schemas"]["LatentsOutput"]; + sd3_l2i: components["schemas"]["ImageOutput"]; + sd3_model_loader: components["schemas"]["Sd3ModelLoaderOutput"]; + sd3_text_encoder: components["schemas"]["SD3ConditioningOutput"]; + sdxl_compel_prompt: components["schemas"]["ConditioningOutput"]; + sdxl_lora_collection_loader: components["schemas"]["SDXLLoRALoaderOutput"]; + sdxl_lora_loader: components["schemas"]["SDXLLoRALoaderOutput"]; + sdxl_model_loader: components["schemas"]["SDXLModelLoaderOutput"]; + sdxl_refiner_compel_prompt: components["schemas"]["ConditioningOutput"]; + sdxl_refiner_model_loader: components["schemas"]["SDXLRefinerModelLoaderOutput"]; + seamless: components["schemas"]["SeamlessModeOutput"]; + segment_anything: components["schemas"]["MaskOutput"]; + show_image: components["schemas"]["ImageOutput"]; + spandrel_image_to_image: components["schemas"]["ImageOutput"]; + spandrel_image_to_image_autoscale: components["schemas"]["ImageOutput"]; + string: components["schemas"]["StringOutput"]; + string_batch: components["schemas"]["StringOutput"]; + string_collection: components["schemas"]["StringCollectionOutput"]; + string_generator: components["schemas"]["StringGeneratorOutput"]; + string_join: components["schemas"]["StringOutput"]; + string_join_three: components["schemas"]["StringOutput"]; + string_replace: components["schemas"]["StringOutput"]; + string_split: components["schemas"]["String2Output"]; + string_split_neg: components["schemas"]["StringPosNegOutput"]; + sub: components["schemas"]["IntegerOutput"]; + t2i_adapter: components["schemas"]["T2IAdapterOutput"]; + tensor_mask_to_image: components["schemas"]["ImageOutput"]; + tile_to_properties: components["schemas"]["TileToPropertiesOutput"]; + tiled_multi_diffusion_denoise_latents: components["schemas"]["LatentsOutput"]; + tomask: components["schemas"]["ImageOutput"]; + unsharp_mask: components["schemas"]["ImageOutput"]; + vae_loader: components["schemas"]["VAEOutput"]; + z_image_control: components["schemas"]["ZImageControlOutput"]; + z_image_denoise: components["schemas"]["LatentsOutput"]; + z_image_denoise_meta: components["schemas"]["LatentsMetaOutput"]; + z_image_i2l: components["schemas"]["LatentsOutput"]; + z_image_l2i: components["schemas"]["ImageOutput"]; + z_image_lora_collection_loader: components["schemas"]["ZImageLoRALoaderOutput"]; + z_image_lora_loader: components["schemas"]["ZImageLoRALoaderOutput"]; + z_image_model_loader: components["schemas"]["ZImageModelLoaderOutput"]; + z_image_seed_variance_enhancer: components["schemas"]["ZImageConditioningOutput"]; + z_image_text_encoder: components["schemas"]["ZImageConditioningOutput"]; }; /** - * Merge String Collections - * @description Merges two Collections of LoRAs into a single Collection. + * InvocationProgressEvent + * @description Event model for invocation_progress */ - MergeStringCollectionsInvocation: { + InvocationProgressEvent: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Timestamp + * @description The timestamp of the event */ - id: string; + timestamp: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Queue Id + * @description The ID of the queue */ - is_intermediate?: boolean; + queue_id: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Item Id + * @description The ID of the queue item */ - use_cache?: boolean; + item_id: number; /** - * Collection 1 - * @description A collection of strings or a single string. - * @default null + * Batch Id + * @description The ID of the queue batch */ - collection1?: string | string[] | null; + batch_id: string; /** - * Collection 2 - * @description A collection of strings or a single string. + * Origin + * @description The origin of the queue item * @default null */ - collection2?: string | string[] | null; + origin: string | null; /** - * type - * @default merge_string_collections_invocation - * @constant + * Destination + * @description The destination of the queue item + * @default null */ - type: "merge_string_collections_invocation"; - }; - /** - * MergeStringCollectionsOutput - * @description Merge String Collections output - */ - MergeStringCollectionsOutput: { + destination: string | null; /** - * Collection - * @description The merged collection + * User Id + * @description The ID of the user who created the queue item + * @default system */ - collection: string[]; + user_id: string; /** - * type - * @default merge_string_collections_output - * @constant + * Session Id + * @description The ID of the session (aka graph execution state) */ - type: "merge_string_collections_output"; - }; - /** - * Merge Tiles to Image - * @description Merge multiple tile images into a single image. - */ - MergeTilesToImageInvocation: { + session_id: string; /** - * @description The board to save the image to - * @default null + * Invocation + * @description The ID of the invocation */ - board?: components["schemas"]["BoardField"] | null; + invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; /** - * @description Optional metadata to be saved with the image - * @default null + * Invocation Source Id + * @description The ID of the prepared invocation's source node */ - metadata?: components["schemas"]["MetadataField"] | null; + invocation_source_id: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Message + * @description A message to display */ - id: string; + message: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Percentage + * @description The percentage of the progress (omit to indicate indeterminate progress) + * @default null */ - is_intermediate?: boolean; + percentage: number | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * @description An image representing the current state of the progress + * @default null */ - use_cache?: boolean; + image: components["schemas"]["ProgressImage"] | null; + }; + /** + * InvocationStartedEvent + * @description Event model for invocation_started + */ + InvocationStartedEvent: { /** - * Tiles With Images - * @description A list of tile images with tile properties. - * @default null + * Timestamp + * @description The timestamp of the event */ - tiles_with_images?: components["schemas"]["TileWithImage"][] | null; + timestamp: number; /** - * Blend Mode - * @description blending type Linear or Seam - * @default Seam - * @enum {string} + * Queue Id + * @description The ID of the queue */ - blend_mode?: "Linear" | "Seam"; + queue_id: string; /** - * Blend Amount - * @description The amount to blend adjacent tiles in pixels. Must be <= the amount of overlap between adjacent tiles. - * @default 32 + * Item Id + * @description The ID of the queue item */ - blend_amount?: number; + item_id: number; /** - * type - * @default merge_tiles_to_image - * @constant + * Batch Id + * @description The ID of the queue batch */ - type: "merge_tiles_to_image"; - }; - /** - * MetadataField - * @description Pydantic model for metadata with custom root of type dict[str, Any]. - * Metadata is stored without a strict schema. - */ - MetadataField: Record; - /** - * Metadata Field Extractor - * @description Extracts the text value from an image's metadata given a key. - * Raises an error if the image has no metadata or if the value is not a string (nesting not permitted). - */ - MetadataFieldExtractorInvocation: { + batch_id: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Origin + * @description The origin of the queue item + * @default null */ - id: string; + origin: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Destination + * @description The destination of the queue item + * @default null */ - is_intermediate?: boolean; + destination: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * User Id + * @description The ID of the user who created the queue item + * @default system */ - use_cache?: boolean; + user_id: string; /** - * @description The image to extract metadata from - * @default null + * Session Id + * @description The ID of the session (aka graph execution state) */ - image?: components["schemas"]["ImageField"] | null; + session_id: string; /** - * Key - * @description The key in the image's metadata to extract the value from - * @default null + * Invocation + * @description The ID of the invocation */ - key?: string | null; + invocation: components["schemas"]["AddInvocation"] | components["schemas"]["AlphaMaskToTensorInvocation"] | components["schemas"]["ApplyMaskTensorToImageInvocation"] | components["schemas"]["ApplyMaskToImageInvocation"] | components["schemas"]["BlankImageInvocation"] | components["schemas"]["BlendLatentsInvocation"] | components["schemas"]["BooleanCollectionInvocation"] | components["schemas"]["BooleanInvocation"] | components["schemas"]["BoundingBoxInvocation"] | components["schemas"]["CLIPSkipInvocation"] | components["schemas"]["CV2InfillInvocation"] | components["schemas"]["CalculateImageTilesEvenSplitInvocation"] | components["schemas"]["CalculateImageTilesInvocation"] | components["schemas"]["CalculateImageTilesMinimumOverlapInvocation"] | components["schemas"]["CannyEdgeDetectionInvocation"] | components["schemas"]["CanvasPasteBackInvocation"] | components["schemas"]["CanvasV2MaskAndCropInvocation"] | components["schemas"]["CenterPadCropInvocation"] | components["schemas"]["CogView4DenoiseInvocation"] | components["schemas"]["CogView4ImageToLatentsInvocation"] | components["schemas"]["CogView4LatentsToImageInvocation"] | components["schemas"]["CogView4ModelLoaderInvocation"] | components["schemas"]["CogView4TextEncoderInvocation"] | components["schemas"]["CollectInvocation"] | components["schemas"]["ColorCorrectInvocation"] | components["schemas"]["ColorInvocation"] | components["schemas"]["ColorMapInvocation"] | components["schemas"]["CompelInvocation"] | components["schemas"]["ConditioningCollectionInvocation"] | components["schemas"]["ConditioningInvocation"] | components["schemas"]["ContentShuffleInvocation"] | components["schemas"]["ControlNetInvocation"] | components["schemas"]["CoreMetadataInvocation"] | components["schemas"]["CreateDenoiseMaskInvocation"] | components["schemas"]["CreateGradientMaskInvocation"] | components["schemas"]["CropImageToBoundingBoxInvocation"] | components["schemas"]["CropLatentsCoreInvocation"] | components["schemas"]["CvInpaintInvocation"] | components["schemas"]["DWOpenposeDetectionInvocation"] | components["schemas"]["DecodeInvisibleWatermarkInvocation"] | components["schemas"]["DenoiseLatentsInvocation"] | components["schemas"]["DenoiseLatentsMetaInvocation"] | components["schemas"]["DepthAnythingDepthEstimationInvocation"] | components["schemas"]["DivideInvocation"] | components["schemas"]["DynamicPromptInvocation"] | components["schemas"]["ESRGANInvocation"] | components["schemas"]["ExpandMaskWithFadeInvocation"] | components["schemas"]["FLUXLoRACollectionLoader"] | components["schemas"]["FaceIdentifierInvocation"] | components["schemas"]["FaceMaskInvocation"] | components["schemas"]["FaceOffInvocation"] | components["schemas"]["FloatBatchInvocation"] | components["schemas"]["FloatCollectionInvocation"] | components["schemas"]["FloatGenerator"] | components["schemas"]["FloatInvocation"] | components["schemas"]["FloatLinearRangeInvocation"] | components["schemas"]["FloatMathInvocation"] | components["schemas"]["FloatToIntegerInvocation"] | components["schemas"]["Flux2DenoiseInvocation"] | components["schemas"]["Flux2KleinLoRACollectionLoader"] | components["schemas"]["Flux2KleinLoRALoaderInvocation"] | components["schemas"]["Flux2KleinModelLoaderInvocation"] | components["schemas"]["Flux2KleinTextEncoderInvocation"] | components["schemas"]["Flux2VaeDecodeInvocation"] | components["schemas"]["Flux2VaeEncodeInvocation"] | components["schemas"]["FluxControlLoRALoaderInvocation"] | components["schemas"]["FluxControlNetInvocation"] | components["schemas"]["FluxDenoiseInvocation"] | components["schemas"]["FluxDenoiseLatentsMetaInvocation"] | components["schemas"]["FluxFillInvocation"] | components["schemas"]["FluxIPAdapterInvocation"] | components["schemas"]["FluxKontextConcatenateImagesInvocation"] | components["schemas"]["FluxKontextInvocation"] | components["schemas"]["FluxLoRALoaderInvocation"] | components["schemas"]["FluxModelLoaderInvocation"] | components["schemas"]["FluxReduxInvocation"] | components["schemas"]["FluxTextEncoderInvocation"] | components["schemas"]["FluxVaeDecodeInvocation"] | components["schemas"]["FluxVaeEncodeInvocation"] | components["schemas"]["FreeUInvocation"] | components["schemas"]["GetMaskBoundingBoxInvocation"] | components["schemas"]["GroundingDinoInvocation"] | components["schemas"]["HEDEdgeDetectionInvocation"] | components["schemas"]["HeuristicResizeInvocation"] | components["schemas"]["IPAdapterInvocation"] | components["schemas"]["IdealSizeInvocation"] | components["schemas"]["ImageBatchInvocation"] | components["schemas"]["ImageBlurInvocation"] | components["schemas"]["ImageChannelInvocation"] | components["schemas"]["ImageChannelMultiplyInvocation"] | components["schemas"]["ImageChannelOffsetInvocation"] | components["schemas"]["ImageCollectionInvocation"] | components["schemas"]["ImageConvertInvocation"] | components["schemas"]["ImageCropInvocation"] | components["schemas"]["ImageGenerator"] | components["schemas"]["ImageHueAdjustmentInvocation"] | components["schemas"]["ImageInverseLerpInvocation"] | components["schemas"]["ImageInvocation"] | components["schemas"]["ImageLerpInvocation"] | components["schemas"]["ImageMaskToTensorInvocation"] | components["schemas"]["ImageMultiplyInvocation"] | components["schemas"]["ImageNSFWBlurInvocation"] | components["schemas"]["ImageNoiseInvocation"] | components["schemas"]["ImagePanelLayoutInvocation"] | components["schemas"]["ImagePasteInvocation"] | components["schemas"]["ImageResizeInvocation"] | components["schemas"]["ImageScaleInvocation"] | components["schemas"]["ImageToLatentsInvocation"] | components["schemas"]["ImageWatermarkInvocation"] | components["schemas"]["InfillColorInvocation"] | components["schemas"]["InfillPatchMatchInvocation"] | components["schemas"]["InfillTileInvocation"] | components["schemas"]["IntegerBatchInvocation"] | components["schemas"]["IntegerCollectionInvocation"] | components["schemas"]["IntegerGenerator"] | components["schemas"]["IntegerInvocation"] | components["schemas"]["IntegerMathInvocation"] | components["schemas"]["InvertTensorMaskInvocation"] | components["schemas"]["InvokeAdjustImageHuePlusInvocation"] | components["schemas"]["InvokeEquivalentAchromaticLightnessInvocation"] | components["schemas"]["InvokeImageBlendInvocation"] | components["schemas"]["InvokeImageCompositorInvocation"] | components["schemas"]["InvokeImageDilateOrErodeInvocation"] | components["schemas"]["InvokeImageEnhanceInvocation"] | components["schemas"]["InvokeImageValueThresholdsInvocation"] | components["schemas"]["IterateInvocation"] | components["schemas"]["LaMaInfillInvocation"] | components["schemas"]["LatentsCollectionInvocation"] | components["schemas"]["LatentsInvocation"] | components["schemas"]["LatentsToImageInvocation"] | components["schemas"]["LineartAnimeEdgeDetectionInvocation"] | components["schemas"]["LineartEdgeDetectionInvocation"] | components["schemas"]["LlavaOnevisionVllmInvocation"] | components["schemas"]["LoRACollectionLoader"] | components["schemas"]["LoRALoaderInvocation"] | components["schemas"]["LoRASelectorInvocation"] | components["schemas"]["MLSDDetectionInvocation"] | components["schemas"]["MainModelLoaderInvocation"] | components["schemas"]["MaskCombineInvocation"] | components["schemas"]["MaskEdgeInvocation"] | components["schemas"]["MaskFromAlphaInvocation"] | components["schemas"]["MaskFromIDInvocation"] | components["schemas"]["MaskTensorToImageInvocation"] | components["schemas"]["MediaPipeFaceDetectionInvocation"] | components["schemas"]["MergeMetadataInvocation"] | components["schemas"]["MergeTilesToImageInvocation"] | components["schemas"]["MetadataFieldExtractorInvocation"] | components["schemas"]["MetadataFromImageInvocation"] | components["schemas"]["MetadataInvocation"] | components["schemas"]["MetadataItemInvocation"] | components["schemas"]["MetadataItemLinkedInvocation"] | components["schemas"]["MetadataToBoolCollectionInvocation"] | components["schemas"]["MetadataToBoolInvocation"] | components["schemas"]["MetadataToControlnetsInvocation"] | components["schemas"]["MetadataToFloatCollectionInvocation"] | components["schemas"]["MetadataToFloatInvocation"] | components["schemas"]["MetadataToIPAdaptersInvocation"] | components["schemas"]["MetadataToIntegerCollectionInvocation"] | components["schemas"]["MetadataToIntegerInvocation"] | components["schemas"]["MetadataToLorasCollectionInvocation"] | components["schemas"]["MetadataToLorasInvocation"] | components["schemas"]["MetadataToModelInvocation"] | components["schemas"]["MetadataToSDXLLorasInvocation"] | components["schemas"]["MetadataToSDXLModelInvocation"] | components["schemas"]["MetadataToSchedulerInvocation"] | components["schemas"]["MetadataToStringCollectionInvocation"] | components["schemas"]["MetadataToStringInvocation"] | components["schemas"]["MetadataToT2IAdaptersInvocation"] | components["schemas"]["MetadataToVAEInvocation"] | components["schemas"]["ModelIdentifierInvocation"] | components["schemas"]["MultiplyInvocation"] | components["schemas"]["NoiseInvocation"] | components["schemas"]["NormalMapInvocation"] | components["schemas"]["PBRMapsInvocation"] | components["schemas"]["PairTileImageInvocation"] | components["schemas"]["PasteImageIntoBoundingBoxInvocation"] | components["schemas"]["PiDiNetEdgeDetectionInvocation"] | components["schemas"]["PromptTemplateInvocation"] | components["schemas"]["PromptsFromFileInvocation"] | components["schemas"]["RandomFloatInvocation"] | components["schemas"]["RandomIntInvocation"] | components["schemas"]["RandomRangeInvocation"] | components["schemas"]["RangeInvocation"] | components["schemas"]["RangeOfSizeInvocation"] | components["schemas"]["RectangleMaskInvocation"] | components["schemas"]["ResizeLatentsInvocation"] | components["schemas"]["RoundInvocation"] | components["schemas"]["SD3DenoiseInvocation"] | components["schemas"]["SD3ImageToLatentsInvocation"] | components["schemas"]["SD3LatentsToImageInvocation"] | components["schemas"]["SDXLCompelPromptInvocation"] | components["schemas"]["SDXLLoRACollectionLoader"] | components["schemas"]["SDXLLoRALoaderInvocation"] | components["schemas"]["SDXLModelLoaderInvocation"] | components["schemas"]["SDXLRefinerCompelPromptInvocation"] | components["schemas"]["SDXLRefinerModelLoaderInvocation"] | components["schemas"]["SaveImageInvocation"] | components["schemas"]["ScaleLatentsInvocation"] | components["schemas"]["SchedulerInvocation"] | components["schemas"]["Sd3ModelLoaderInvocation"] | components["schemas"]["Sd3TextEncoderInvocation"] | components["schemas"]["SeamlessModeInvocation"] | components["schemas"]["SegmentAnythingInvocation"] | components["schemas"]["ShowImageInvocation"] | components["schemas"]["SpandrelImageToImageAutoscaleInvocation"] | components["schemas"]["SpandrelImageToImageInvocation"] | components["schemas"]["StringBatchInvocation"] | components["schemas"]["StringCollectionInvocation"] | components["schemas"]["StringGenerator"] | components["schemas"]["StringInvocation"] | components["schemas"]["StringJoinInvocation"] | components["schemas"]["StringJoinThreeInvocation"] | components["schemas"]["StringReplaceInvocation"] | components["schemas"]["StringSplitInvocation"] | components["schemas"]["StringSplitNegInvocation"] | components["schemas"]["SubtractInvocation"] | components["schemas"]["T2IAdapterInvocation"] | components["schemas"]["TileToPropertiesInvocation"] | components["schemas"]["TiledMultiDiffusionDenoiseLatents"] | components["schemas"]["UnsharpMaskInvocation"] | components["schemas"]["VAELoaderInvocation"] | components["schemas"]["ZImageControlInvocation"] | components["schemas"]["ZImageDenoiseInvocation"] | components["schemas"]["ZImageDenoiseMetaInvocation"] | components["schemas"]["ZImageImageToLatentsInvocation"] | components["schemas"]["ZImageLatentsToImageInvocation"] | components["schemas"]["ZImageLoRACollectionLoader"] | components["schemas"]["ZImageLoRALoaderInvocation"] | components["schemas"]["ZImageModelLoaderInvocation"] | components["schemas"]["ZImageSeedVarianceEnhancerInvocation"] | components["schemas"]["ZImageTextEncoderInvocation"]; /** - * type - * @default metadata_field_extractor - * @constant + * Invocation Source Id + * @description The ID of the prepared invocation's source node */ - type: "metadata_field_extractor"; + invocation_source_id: string; }; /** - * Metadata From Image - * @description Used to create a core metadata item then Add/Update it to the provided metadata + * InvokeAIAppConfig + * @description Invoke's global app configuration. + * + * Typically, you won't need to interact with this class directly. Instead, use the `get_config` function from `invokeai.app.services.config` to get a singleton config object. + * + * Attributes: + * host: IP address to bind to. Use `0.0.0.0` to serve to your local network. + * port: Port to bind to. + * allow_origins: Allowed CORS origins. + * allow_credentials: Allow CORS credentials. + * allow_methods: Methods allowed for CORS. + * allow_headers: Headers allowed for CORS. + * ssl_certfile: SSL certificate file for HTTPS. See https://www.uvicorn.org/settings/#https. + * ssl_keyfile: SSL key file for HTTPS. See https://www.uvicorn.org/settings/#https. + * log_tokenization: Enable logging of parsed prompt tokens. + * patchmatch: Enable patchmatch inpaint code. + * models_dir: Path to the models directory. + * convert_cache_dir: Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). + * download_cache_dir: Path to the directory that contains dynamically downloaded models. + * legacy_conf_dir: Path to directory of legacy checkpoint config files. + * db_dir: Path to InvokeAI databases directory. + * outputs_dir: Path to directory for outputs. + * custom_nodes_dir: Path to directory for custom nodes. + * style_presets_dir: Path to directory for style presets. + * workflow_thumbnails_dir: Path to directory for workflow thumbnails. + * log_handlers: Log handler. Valid options are "console", "file=", "syslog=path|address:host:port", "http=". + * log_format: Log format. Use "plain" for text-only, "color" for colorized output, "legacy" for 2.3-style logging and "syslog" for syslog-style.
Valid values: `plain`, `color`, `syslog`, `legacy` + * log_level: Emit logging messages at this level or higher.
Valid values: `debug`, `info`, `warning`, `error`, `critical` + * log_sql: Log SQL queries. `log_level` must be `debug` for this to do anything. Extremely verbose. + * log_level_network: Log level for network-related messages. 'info' and 'debug' are very verbose.
Valid values: `debug`, `info`, `warning`, `error`, `critical` + * use_memory_db: Use in-memory database. Useful for development. + * dev_reload: Automatically reload when Python sources are changed. Does not reload node definitions. + * profile_graphs: Enable graph profiling using `cProfile`. + * profile_prefix: An optional prefix for profile output files. + * profiles_dir: Path to profiles output directory. + * max_cache_ram_gb: The maximum amount of CPU RAM to use for model caching in GB. If unset, the limit will be configured based on the available RAM. In most cases, it is recommended to leave this unset. + * max_cache_vram_gb: The amount of VRAM to use for model caching in GB. If unset, the limit will be configured based on the available VRAM and the device_working_mem_gb. In most cases, it is recommended to leave this unset. + * log_memory_usage: If True, a memory snapshot will be captured before and after every model cache operation, and the result will be logged (at debug level). There is a time cost to capturing the memory snapshots, so it is recommended to only enable this feature if you are actively inspecting the model cache's behaviour. + * model_cache_keep_alive_min: How long to keep models in cache after last use, in minutes. A value of 0 (the default) means models are kept in cache indefinitely. If no model generations occur within the timeout period, the model cache is cleared using the same logic as the 'Clear Model Cache' button. + * device_working_mem_gb: The amount of working memory to keep available on the compute device (in GB). Has no effect if running on CPU. If you are experiencing OOM errors, try increasing this value. + * enable_partial_loading: Enable partial loading of models. This enables models to run with reduced VRAM requirements (at the cost of slower speed) by streaming the model from RAM to VRAM as its used. In some edge cases, partial loading can cause models to run more slowly if they were previously being fully loaded into VRAM. + * keep_ram_copy_of_weights: Whether to keep a full RAM copy of a model's weights when the model is loaded in VRAM. Keeping a RAM copy increases average RAM usage, but speeds up model switching and LoRA patching (assuming there is sufficient RAM). Set this to False if RAM pressure is consistently high. + * ram: DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_ram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. + * vram: DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_vram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. + * lazy_offload: DEPRECATED: This setting is no longer used. Lazy-offloading is enabled by default. This config setting will be removed once the new model cache behavior is stable. + * pytorch_cuda_alloc_conf: Configure the Torch CUDA memory allocator. This will impact peak reserved VRAM usage and performance. Setting to "backend:cudaMallocAsync" works well on many systems. The optimal configuration is highly dependent on the system configuration (device type, VRAM, CUDA driver version, etc.), so must be tuned experimentally. + * device: Preferred execution device. `auto` will choose the device depending on the hardware platform and the installed torch capabilities.
Valid values: `auto`, `cpu`, `cuda`, `mps`, `cuda:N` (where N is a device number) + * precision: Floating point precision. `float16` will consume half the memory of `float32` but produce slightly lower-quality images. The `auto` setting will guess the proper precision based on your video card and operating system.
Valid values: `auto`, `float16`, `bfloat16`, `float32` + * sequential_guidance: Whether to calculate guidance in serial instead of in parallel, lowering memory requirements. + * attention_type: Attention type.
Valid values: `auto`, `normal`, `xformers`, `sliced`, `torch-sdp` + * attention_slice_size: Slice size, valid when attention_type=="sliced".
Valid values: `auto`, `balanced`, `max`, `1`, `2`, `3`, `4`, `5`, `6`, `7`, `8` + * force_tiled_decode: Whether to enable tiled VAE decode (reduces memory consumption with some performance penalty). + * pil_compress_level: The compress_level setting of PIL.Image.save(), used for PNG encoding. All settings are lossless. 0 = no compression, 1 = fastest with slightly larger filesize, 9 = slowest with smallest filesize. 1 is typically the best setting. + * max_queue_size: Maximum number of items in the session queue. + * clear_queue_on_startup: Empties session queue on startup. + * allow_nodes: List of nodes to allow. Omit to allow all. + * deny_nodes: List of nodes to deny. Omit to deny none. + * node_cache_size: How many cached nodes to keep in memory. + * hashing_algorithm: Model hashing algorthim for model installs. 'blake3_multi' is best for SSDs. 'blake3_single' is best for spinning disk HDDs. 'random' disables hashing, instead assigning a UUID to models. Useful when using a memory db to reduce model installation time, or if you don't care about storing stable hashes for models. Alternatively, any other hashlib algorithm is accepted, though these are not nearly as performant as blake3.
Valid values: `blake3_multi`, `blake3_single`, `random`, `md5`, `sha1`, `sha224`, `sha256`, `sha384`, `sha512`, `blake2b`, `blake2s`, `sha3_224`, `sha3_256`, `sha3_384`, `sha3_512`, `shake_128`, `shake_256` + * remote_api_tokens: List of regular expression and token pairs used when downloading models from URLs. The download URL is tested against the regex, and if it matches, the token is provided in as a Bearer token. + * scan_models_on_startup: Scan the models directory on startup, registering orphaned models. This is typically only used in conjunction with `use_memory_db` for testing purposes. + * unsafe_disable_picklescan: UNSAFE. Disable the picklescan security check during model installation. Recommended only for development and testing purposes. This will allow arbitrary code execution during model installation, so should never be used in production. + * allow_unknown_models: Allow installation of models that we are unable to identify. If enabled, models will be marked as `unknown` in the database, and will not have any metadata associated with them. If disabled, unknown models will be rejected during installation. + * multiuser: Enable multiuser support. When disabled, the application runs in single-user mode using a default system account with administrator privileges. When enabled, requires user authentication and authorization. + * strict_password_checking: Enforce strict password requirements. When True, passwords must contain uppercase, lowercase, and numbers. When False (default), any password is accepted but its strength (weak/moderate/strong) is reported to the user. */ - MetadataFromImageInvocation: { + InvokeAIAppConfig: { + /** + * Schema Version + * @description Schema version of the config file. This is not a user-configurable setting. + * @default 4.0.2 + */ + schema_version?: string; + /** + * Legacy Models Yaml Path + * @description Path to the legacy models.yaml file. This is not a user-configurable setting. + */ + legacy_models_yaml_path?: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Host + * @description IP address to bind to. Use `0.0.0.0` to serve to your local network. + * @default 127.0.0.1 */ - id: string; + host?: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Port + * @description Port to bind to. + * @default 9090 */ - is_intermediate?: boolean; + port?: number; /** - * Use Cache - * @description Whether or not to use the cache + * Allow Origins + * @description Allowed CORS origins. + * @default [] + */ + allow_origins?: string[]; + /** + * Allow Credentials + * @description Allow CORS credentials. * @default true */ - use_cache?: boolean; + allow_credentials?: boolean; /** - * @description The image to process - * @default null + * Allow Methods + * @description Methods allowed for CORS. + * @default [ + * "*" + * ] */ - image?: components["schemas"]["ImageField"] | null; + allow_methods?: string[]; /** - * type - * @default metadata_from_image - * @constant + * Allow Headers + * @description Headers allowed for CORS. + * @default [ + * "*" + * ] */ - type: "metadata_from_image"; - }; - /** - * Metadata - * @description Takes a MetadataItem or collection of MetadataItems and outputs a MetadataDict. - */ - MetadataInvocation: { + allow_headers?: string[]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Ssl Certfile + * @description SSL certificate file for HTTPS. See https://www.uvicorn.org/settings/#https. */ - id: string; + ssl_certfile?: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Ssl Keyfile + * @description SSL key file for HTTPS. See https://www.uvicorn.org/settings/#https. + */ + ssl_keyfile?: string | null; + /** + * Log Tokenization + * @description Enable logging of parsed prompt tokens. * @default false */ - is_intermediate?: boolean; + log_tokenization?: boolean; /** - * Use Cache - * @description Whether or not to use the cache + * Patchmatch + * @description Enable patchmatch inpaint code. * @default true */ - use_cache?: boolean; + patchmatch?: boolean; /** - * Items - * @description A single metadata item or collection of metadata items - * @default null + * Models Dir + * Format: path + * @description Path to the models directory. + * @default models */ - items?: components["schemas"]["MetadataItemField"][] | components["schemas"]["MetadataItemField"] | null; + models_dir?: string; /** - * type - * @default metadata - * @constant + * Convert Cache Dir + * Format: path + * @description Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). + * @default models\.convert_cache */ - type: "metadata"; - }; - /** MetadataItemField */ - MetadataItemField: { + convert_cache_dir?: string; /** - * Label - * @description Label for this metadata item + * Download Cache Dir + * Format: path + * @description Path to the directory that contains dynamically downloaded models. + * @default models\.download_cache */ - label: string; + download_cache_dir?: string; /** - * Value - * @description The value for this metadata item (may be any type) + * Legacy Conf Dir + * Format: path + * @description Path to directory of legacy checkpoint config files. + * @default configs */ - value: unknown; - }; - /** - * Metadata Item - * @description Used to create an arbitrary metadata item. Provide "label" and make a connection to "value" to store that data as the value. - */ - MetadataItemInvocation: { + legacy_conf_dir?: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Db Dir + * Format: path + * @description Path to InvokeAI databases directory. + * @default databases */ - id: string; + db_dir?: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Outputs Dir + * Format: path + * @description Path to directory for outputs. + * @default outputs */ - is_intermediate?: boolean; + outputs_dir?: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Custom Nodes Dir + * Format: path + * @description Path to directory for custom nodes. + * @default nodes */ - use_cache?: boolean; + custom_nodes_dir?: string; /** - * Label - * @description Label for this metadata item - * @default null + * Style Presets Dir + * Format: path + * @description Path to directory for style presets. + * @default style_presets */ - label?: string | null; + style_presets_dir?: string; /** - * Value - * @description The value for this metadata item (may be any type) - * @default null + * Workflow Thumbnails Dir + * Format: path + * @description Path to directory for workflow thumbnails. + * @default workflow_thumbnails */ - value?: unknown | null; + workflow_thumbnails_dir?: string; /** - * type - * @default metadata_item - * @constant + * Log Handlers + * @description Log handler. Valid options are "console", "file=", "syslog=path|address:host:port", "http=". + * @default [ + * "console" + * ] */ - type: "metadata_item"; - }; - /** - * Metadata Item Linked - * @description Used to Create/Add/Update a value into a metadata label - */ - MetadataItemLinkedInvocation: { + log_handlers?: string[]; /** - * @description Optional metadata to be saved with the image - * @default null + * Log Format + * @description Log format. Use "plain" for text-only, "color" for colorized output, "legacy" for 2.3-style logging and "syslog" for syslog-style. + * @default color + * @enum {string} */ - metadata?: components["schemas"]["MetadataField"] | null; + log_format?: "plain" | "color" | "syslog" | "legacy"; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Log Level + * @description Emit logging messages at this level or higher. + * @default info + * @enum {string} */ - id: string; + log_level?: "debug" | "info" | "warning" | "error" | "critical"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Log Sql + * @description Log SQL queries. `log_level` must be `debug` for this to do anything. Extremely verbose. * @default false */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; + log_sql?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * + * Log Level Network + * @description Log level for network-related messages. 'info' and 'debug' are very verbose. + * @default warning * @enum {string} */ - label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt" | "width" | "height" | "seed" | "cfg_scale" | "cfg_rescale_multiplier" | "steps" | "scheduler" | "clip_skip" | "model" | "vae" | "seamless_x" | "seamless_y" | "guidance" | "cfg_scale_start_step" | "cfg_scale_end_step"; + log_level_network?: "debug" | "info" | "warning" | "error" | "critical"; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Use Memory Db + * @description Use in-memory database. Useful for development. + * @default false */ - custom_label?: string | null; + use_memory_db?: boolean; /** - * Value - * @description The value for this metadata item (may be any type) - * @default null + * Dev Reload + * @description Automatically reload when Python sources are changed. Does not reload node definitions. + * @default false */ - value?: unknown | null; + dev_reload?: boolean; /** - * type - * @default metadata_item_linked - * @constant + * Profile Graphs + * @description Enable graph profiling using `cProfile`. + * @default false */ - type: "metadata_item_linked"; - }; - /** - * MetadataItemOutput - * @description Metadata Item Output - */ - MetadataItemOutput: { - /** @description Metadata Item */ - item: components["schemas"]["MetadataItemField"]; + profile_graphs?: boolean; /** - * type - * @default metadata_item_output - * @constant - */ - type: "metadata_item_output"; - }; - /** MetadataOutput */ - MetadataOutput: { - /** @description Metadata Dict */ - metadata: components["schemas"]["MetadataField"]; + * Profile Prefix + * @description An optional prefix for profile output files. + */ + profile_prefix?: string | null; /** - * type - * @default metadata_output - * @constant + * Profiles Dir + * Format: path + * @description Path to profiles output directory. + * @default profiles */ - type: "metadata_output"; - }; - /** - * Metadata To Bool Collection - * @description Extracts a Boolean value Collection of a label from metadata - */ - MetadataToBoolCollectionInvocation: { + profiles_dir?: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Max Cache Ram Gb + * @description The maximum amount of CPU RAM to use for model caching in GB. If unset, the limit will be configured based on the available RAM. In most cases, it is recommended to leave this unset. */ - metadata?: components["schemas"]["MetadataField"] | null; + max_cache_ram_gb?: number | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Max Cache Vram Gb + * @description The amount of VRAM to use for model caching in GB. If unset, the limit will be configured based on the available VRAM and the device_working_mem_gb. In most cases, it is recommended to leave this unset. */ - id: string; + max_cache_vram_gb?: number | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Log Memory Usage + * @description If True, a memory snapshot will be captured before and after every model cache operation, and the result will be logged (at debug level). There is a time cost to capturing the memory snapshots, so it is recommended to only enable this feature if you are actively inspecting the model cache's behaviour. * @default false */ - is_intermediate?: boolean; + log_memory_usage?: boolean; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Model Cache Keep Alive Min + * @description How long to keep models in cache after last use, in minutes. A value of 0 (the default) means models are kept in cache indefinitely. If no model generations occur within the timeout period, the model cache is cleared using the same logic as the 'Clear Model Cache' button. + * @default 0 */ - use_cache?: boolean; + model_cache_keep_alive_min?: number; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Device Working Mem Gb + * @description The amount of working memory to keep available on the compute device (in GB). Has no effect if running on CPU. If you are experiencing OOM errors, try increasing this value. + * @default 3 */ - label?: "* CUSTOM LABEL *" | "seamless_x" | "seamless_y"; + device_working_mem_gb?: number; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Enable Partial Loading + * @description Enable partial loading of models. This enables models to run with reduced VRAM requirements (at the cost of slower speed) by streaming the model from RAM to VRAM as its used. In some edge cases, partial loading can cause models to run more slowly if they were previously being fully loaded into VRAM. + * @default false */ - custom_label?: string | null; + enable_partial_loading?: boolean; /** - * Default Value - * @description The default bool to use if not found in the metadata - * @default null + * Keep Ram Copy Of Weights + * @description Whether to keep a full RAM copy of a model's weights when the model is loaded in VRAM. Keeping a RAM copy increases average RAM usage, but speeds up model switching and LoRA patching (assuming there is sufficient RAM). Set this to False if RAM pressure is consistently high. + * @default true */ - default_value?: boolean[] | null; + keep_ram_copy_of_weights?: boolean; /** - * type - * @default metadata_to_bool_collection - * @constant + * Ram + * @description DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_ram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. */ - type: "metadata_to_bool_collection"; - }; - /** - * Metadata To Bool - * @description Extracts a Boolean value of a label from metadata - */ - MetadataToBoolInvocation: { + ram?: number | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Vram + * @description DEPRECATED: This setting is no longer used. It has been replaced by `max_cache_vram_gb`, but most users will not need to use this config since automatic cache size limits should work well in most cases. This config setting will be removed once the new model cache behavior is stable. */ - metadata?: components["schemas"]["MetadataField"] | null; + vram?: number | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Lazy Offload + * @description DEPRECATED: This setting is no longer used. Lazy-offloading is enabled by default. This config setting will be removed once the new model cache behavior is stable. + * @default true */ - id: string; + lazy_offload?: boolean; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Pytorch Cuda Alloc Conf + * @description Configure the Torch CUDA memory allocator. This will impact peak reserved VRAM usage and performance. Setting to "backend:cudaMallocAsync" works well on many systems. The optimal configuration is highly dependent on the system configuration (device type, VRAM, CUDA driver version, etc.), so must be tuned experimentally. */ - is_intermediate?: boolean; + pytorch_cuda_alloc_conf?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Device + * @description Preferred execution device. `auto` will choose the device depending on the hardware platform and the installed torch capabilities.
Valid values: `auto`, `cpu`, `cuda`, `mps`, `cuda:N` (where N is a device number) + * @default auto */ - use_cache?: boolean; + device?: string; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * + * Precision + * @description Floating point precision. `float16` will consume half the memory of `float32` but produce slightly lower-quality images. The `auto` setting will guess the proper precision based on your video card and operating system. + * @default auto * @enum {string} */ - label?: "* CUSTOM LABEL *" | "seamless_x" | "seamless_y"; + precision?: "auto" | "float16" | "bfloat16" | "float32"; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Sequential Guidance + * @description Whether to calculate guidance in serial instead of in parallel, lowering memory requirements. + * @default false */ - custom_label?: string | null; + sequential_guidance?: boolean; /** - * Default Value - * @description The default bool to use if not found in the metadata - * @default null + * Attention Type + * @description Attention type. + * @default auto + * @enum {string} */ - default_value?: boolean | null; + attention_type?: "auto" | "normal" | "xformers" | "sliced" | "torch-sdp"; /** - * type - * @default metadata_to_bool - * @constant + * Attention Slice Size + * @description Slice size, valid when attention_type=="sliced". + * @default auto + * @enum {unknown} */ - type: "metadata_to_bool"; - }; - /** - * Metadata To ControlNets - * @description Extracts a Controlnets value of a label from metadata - */ - MetadataToControlnetsInvocation: { + attention_slice_size?: "auto" | "balanced" | "max" | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8; /** - * @description Optional metadata to be saved with the image - * @default null + * Force Tiled Decode + * @description Whether to enable tiled VAE decode (reduces memory consumption with some performance penalty). + * @default false */ - metadata?: components["schemas"]["MetadataField"] | null; + force_tiled_decode?: boolean; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Pil Compress Level + * @description The compress_level setting of PIL.Image.save(), used for PNG encoding. All settings are lossless. 0 = no compression, 1 = fastest with slightly larger filesize, 9 = slowest with smallest filesize. 1 is typically the best setting. + * @default 1 */ - id: string; + pil_compress_level?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Max Queue Size + * @description Maximum number of items in the session queue. + * @default 10000 + */ + max_queue_size?: number; + /** + * Clear Queue On Startup + * @description Empties session queue on startup. * @default false */ - is_intermediate?: boolean; + clear_queue_on_startup?: boolean; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Allow Nodes + * @description List of nodes to allow. Omit to allow all. */ - use_cache?: boolean; + allow_nodes?: string[] | null; /** - * ControlNet-List - * @default null + * Deny Nodes + * @description List of nodes to deny. Omit to deny none. */ - control_list?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + deny_nodes?: string[] | null; /** - * type - * @default metadata_to_controlnets - * @constant + * Node Cache Size + * @description How many cached nodes to keep in memory. + * @default 512 */ - type: "metadata_to_controlnets"; - }; - /** - * Metadata To Float Collection - * @description Extracts a Float value Collection of a label from metadata - */ - MetadataToFloatCollectionInvocation: { + node_cache_size?: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Hashing Algorithm + * @description Model hashing algorthim for model installs. 'blake3_multi' is best for SSDs. 'blake3_single' is best for spinning disk HDDs. 'random' disables hashing, instead assigning a UUID to models. Useful when using a memory db to reduce model installation time, or if you don't care about storing stable hashes for models. Alternatively, any other hashlib algorithm is accepted, though these are not nearly as performant as blake3. + * @default blake3_single + * @enum {string} */ - metadata?: components["schemas"]["MetadataField"] | null; + hashing_algorithm?: "blake3_multi" | "blake3_single" | "random" | "md5" | "sha1" | "sha224" | "sha256" | "sha384" | "sha512" | "blake2b" | "blake2s" | "sha3_224" | "sha3_256" | "sha3_384" | "sha3_512" | "shake_128" | "shake_256"; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Remote Api Tokens + * @description List of regular expression and token pairs used when downloading models from URLs. The download URL is tested against the regex, and if it matches, the token is provided in as a Bearer token. */ - id: string; + remote_api_tokens?: components["schemas"]["URLRegexTokenPair"][] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Scan Models On Startup + * @description Scan the models directory on startup, registering orphaned models. This is typically only used in conjunction with `use_memory_db` for testing purposes. * @default false */ - is_intermediate?: boolean; + scan_models_on_startup?: boolean; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Unsafe Disable Picklescan + * @description UNSAFE. Disable the picklescan security check during model installation. Recommended only for development and testing purposes. This will allow arbitrary code execution during model installation, so should never be used in production. + * @default false */ - use_cache?: boolean; + unsafe_disable_picklescan?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Allow Unknown Models + * @description Allow installation of models that we are unable to identify. If enabled, models will be marked as `unknown` in the database, and will not have any metadata associated with them. If disabled, unknown models will be rejected during installation. + * @default true */ - label?: "* CUSTOM LABEL *" | "cfg_scale" | "cfg_rescale_multiplier" | "guidance"; + allow_unknown_models?: boolean; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Multiuser + * @description Enable multiuser support. When disabled, the application runs in single-user mode using a default system account with administrator privileges. When enabled, requires user authentication and authorization. + * @default false */ - custom_label?: string | null; + multiuser?: boolean; /** - * Default Value - * @description The default float to use if not found in the metadata - * @default null + * Strict Password Checking + * @description Enforce strict password requirements. When True, passwords must contain uppercase, lowercase, and numbers. When False (default), any password is accepted but its strength (weak/moderate/strong) is reported to the user. + * @default false */ - default_value?: number[] | null; + strict_password_checking?: boolean; + }; + /** + * InvokeAIAppConfigWithSetFields + * @description InvokeAI App Config with model fields set + */ + InvokeAIAppConfigWithSetFields: { /** - * type - * @default metadata_to_float_collection - * @constant + * Set Fields + * @description The set fields */ - type: "metadata_to_float_collection"; + set_fields: string[]; + /** @description The InvokeAI App Config */ + config: components["schemas"]["InvokeAIAppConfig"]; }; /** - * Metadata To Float - * @description Extracts a Float value of a label from metadata + * Adjust Image Hue Plus + * @description Adjusts the Hue of an image by rotating it in the selected color space. Originally created by @dwringer */ - MetadataToFloatInvocation: { + InvokeAdjustImageHuePlusInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -29894,36 +14915,58 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * + * @description The image to adjust + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Space + * @description Color space in which to rotate hue by polar coords (*: non-invertible) + * @default HSV / HSL / RGB * @enum {string} */ - label?: "* CUSTOM LABEL *" | "cfg_scale" | "cfg_rescale_multiplier" | "guidance"; + space?: "HSV / HSL / RGB" | "Okhsl" | "Okhsv" | "*Oklch / Oklab" | "*LCh / CIELab" | "*UPLab (w/CIELab_to_UPLab.icc)"; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Degrees + * @description Degrees by which to rotate image hue + * @default 0 */ - custom_label?: string | null; + degrees?: number; /** - * Default Value - * @description The default float to use if not found in the metadata - * @default null + * Preserve Lightness + * @description Whether to preserve CIELAB lightness values + * @default false */ - default_value?: number | null; + preserve_lightness?: boolean; + /** + * Ok Adaptive Gamut + * @description Higher preserves chroma at the expense of lightness (Oklab) + * @default 0.05 + */ + ok_adaptive_gamut?: number; + /** + * Ok High Precision + * @description Use more steps in computing gamut (Oklab/Okhsv/Okhsl) + * @default true + */ + ok_high_precision?: boolean; /** * type - * @default metadata_to_float + * @default invokeai_img_hue_adjust_plus * @constant */ - type: "metadata_to_float"; + type: "invokeai_img_hue_adjust_plus"; }; /** - * Metadata To IP-Adapters - * @description Extracts a IP-Adapters value of a label from metadata + * Equivalent Achromatic Lightness + * @description Calculate Equivalent Achromatic Lightness from image. Originally created by @dwringer */ - MetadataToIPAdaptersInvocation: { + InvokeEquivalentAchromaticLightnessInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -29947,23 +14990,27 @@ export type components = { */ use_cache?: boolean; /** - * IP-Adapter-List - * @description IP-Adapter to apply + * @description Image from which to get channel * @default null */ - ip_adapter_list?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default metadata_to_ip_adapters + * @default invokeai_ealightness * @constant */ - type: "metadata_to_ip_adapters"; + type: "invokeai_ealightness"; }; /** - * Metadata To Integer Collection - * @description Extracts an integer value Collection of a label from metadata + * Image Layer Blend + * @description Blend two images together, with optional opacity, mask, and blend modes. Originally created by @dwringer */ - MetadataToIntegerCollectionInvocation: { + InvokeImageBlendInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -29987,89 +15034,81 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} - */ - label?: "* CUSTOM LABEL *" | "width" | "height" | "seed" | "steps" | "clip_skip" | "cfg_scale_start_step" | "cfg_scale_end_step"; - /** - * Custom Label - * @description Label for this metadata item + * @description The top image to blend * @default null */ - custom_label?: string | null; + layer_upper?: components["schemas"]["ImageField"] | null; /** - * Default Value - * @description The default integer to use if not found in the metadata - * @default null + * Blend Mode + * @description Available blend modes + * @default Normal + * @enum {string} */ - default_value?: number[] | null; + blend_mode?: "Normal" | "Lighten Only" | "Darken Only" | "Lighten Only (EAL)" | "Darken Only (EAL)" | "Hue" | "Saturation" | "Color" | "Luminosity" | "Linear Dodge (Add)" | "Subtract" | "Multiply" | "Divide" | "Screen" | "Overlay" | "Linear Burn" | "Difference" | "Hard Light" | "Soft Light" | "Vivid Light" | "Linear Light" | "Color Burn" | "Color Dodge"; /** - * type - * @default metadata_to_integer_collection - * @constant + * Opacity + * @description Desired opacity of the upper layer + * @default 1 */ - type: "metadata_to_integer_collection"; - }; - /** - * Metadata To Integer - * @description Extracts an integer value of a label from metadata - */ - MetadataToIntegerInvocation: { + opacity?: number; /** - * @description Optional metadata to be saved with the image + * @description Optional mask, used to restrict areas from blending * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + mask?: components["schemas"]["ImageField"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Fit To Width + * @description Scale upper layer to fit base width * @default false */ - is_intermediate?: boolean; + fit_to_width?: boolean; /** - * Use Cache - * @description Whether or not to use the cache + * Fit To Height + * @description Scale upper layer to fit base height * @default true */ - use_cache?: boolean; + fit_to_height?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * + * @description The bottom image to blend + * @default null + */ + layer_base?: components["schemas"]["ImageField"] | null; + /** + * Color Space + * @description Available color spaces for blend computations + * @default RGB * @enum {string} */ - label?: "* CUSTOM LABEL *" | "width" | "height" | "seed" | "steps" | "clip_skip" | "cfg_scale_start_step" | "cfg_scale_end_step"; + color_space?: "RGB" | "Linear RGB" | "HSL (RGB)" | "HSV (RGB)" | "Okhsl" | "Okhsv" | "Oklch (Oklab)" | "LCh (CIELab)"; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Adaptive Gamut + * @description Adaptive gamut clipping (0=off). Higher prioritizes chroma over lightness + * @default 0 */ - custom_label?: string | null; + adaptive_gamut?: number; /** - * Default Value - * @description The default integer to use if not found in the metadata - * @default null + * High Precision + * @description Use more steps in computing gamut when possible + * @default true */ - default_value?: number | null; + high_precision?: boolean; /** * type - * @default metadata_to_integer + * @default invokeai_img_blend * @constant */ - type: "metadata_to_integer"; + type: "invokeai_img_blend"; }; /** - * Metadata To LoRA Collection - * @description Extracts Lora(s) from metadata into a collection + * Image Compositor + * @description Removes backdrop from subject image then overlays subject on background image. Originally created by @dwringer */ - MetadataToLorasCollectionInvocation: { + InvokeImageCompositorInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -30093,92 +15132,63 @@ export type components = { */ use_cache?: boolean; /** - * Custom Label - * @description Label for this metadata item - * @default loras - */ - custom_label?: string; - /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. - * @default [] - */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; - /** - * type - * @default metadata_to_lora_collection - * @constant - */ - type: "metadata_to_lora_collection"; - }; - /** - * MetadataToLorasCollectionOutput - * @description Model loader output - */ - MetadataToLorasCollectionOutput: { - /** - * LoRAs - * @description Collection of LoRA model and weights + * @description Image of the subject on a plain monochrome background + * @default null */ - lora: components["schemas"]["LoRAField"][]; + image_subject?: components["schemas"]["ImageField"] | null; /** - * type - * @default metadata_to_lora_collection_output - * @constant + * @description Image of a background scene + * @default null */ - type: "metadata_to_lora_collection_output"; - }; - /** - * Metadata To LoRAs - * @description Extracts a Loras value of a label from metadata - */ - MetadataToLorasInvocation: { + image_background?: components["schemas"]["ImageField"] | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Chroma Key + * @description Can be empty for corner flood select, or CSS-3 color or tuple + * @default */ - metadata?: components["schemas"]["MetadataField"] | null; + chroma_key?: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Threshold + * @description Subject isolation flood-fill threshold + * @default 50 */ - id: string; + threshold?: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Fill X + * @description Scale base subject image to fit background width * @default false */ - is_intermediate?: boolean; + fill_x?: boolean; /** - * Use Cache - * @description Whether or not to use the cache + * Fill Y + * @description Scale base subject image to fit background height * @default true */ - use_cache?: boolean; + fill_y?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * X Offset + * @description x-offset for the subject + * @default 0 */ - unet?: components["schemas"]["UNetField"] | null; + x_offset?: number; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Y Offset + * @description y-offset for the subject + * @default 0 */ - clip?: components["schemas"]["CLIPField"] | null; + y_offset?: number; /** * type - * @default metadata_to_loras + * @default invokeai_img_composite * @constant */ - type: "metadata_to_loras"; + type: "invokeai_img_composite"; }; /** - * Metadata To Model - * @description Extracts a Model value of a label from metadata + * Image Dilate or Erode + * @description Dilate (expand) or erode (contract) an image. Originally created by @dwringer */ - MetadataToModelInvocation: { + InvokeImageDilateOrErodeInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -30202,72 +15212,52 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default model - * @enum {string} - */ - label?: "* CUSTOM LABEL *" | "model"; - /** - * Custom Label - * @description Label for this metadata item - * @default null - */ - custom_label?: string | null; - /** - * @description The default model to use if not found in the metadata + * @description The image from which to create a mask * @default null */ - default_value?: components["schemas"]["ModelIdentifierField"] | null; - /** - * type - * @default metadata_to_model - * @constant - */ - type: "metadata_to_model"; - }; - /** - * MetadataToModelOutput - * @description String to main model output - */ - MetadataToModelOutput: { - /** - * Model - * @description Main model (UNet, VAE, CLIP) to load - */ - model: components["schemas"]["ModelIdentifierField"]; + image?: components["schemas"]["ImageField"] | null; /** - * Name - * @description Model Name + * Lightness Only + * @description If true, only applies to image lightness (CIELa*b*) + * @default false */ - name: string; + lightness_only?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) + * Radius W + * @description Width (in pixels) by which to dilate(expand) or erode (contract) the image + * @default 4 */ - unet: components["schemas"]["UNetField"]; + radius_w?: number; /** - * VAE - * @description VAE + * Radius H + * @description Height (in pixels) by which to dilate(expand) or erode (contract) the image + * @default 4 */ - vae: components["schemas"]["VAEField"]; + radius_h?: number; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Mode + * @description How to operate on the image + * @default Dilate + * @enum {string} */ - clip: components["schemas"]["CLIPField"]; + mode?: "Dilate" | "Erode"; /** * type - * @default metadata_to_model_output + * @default invokeai_img_dilate_erode * @constant */ - type: "metadata_to_model_output"; + type: "invokeai_img_dilate_erode"; }; /** - * Metadata To SDXL LoRAs - * @description Extracts a SDXL Loras value of a label from metadata + * Enhance Image + * @description Applies processing from PIL's ImageEnhance module. Originally created by @dwringer */ - MetadataToSDXLLorasInvocation: { + InvokeImageEnhanceInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -30291,35 +15281,57 @@ export type components = { */ use_cache?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) + * @description The image for which to apply processing * @default null */ - unet?: components["schemas"]["UNetField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Invert + * @description Whether to invert the image colors + * @default false */ - clip?: components["schemas"]["CLIPField"] | null; + invert?: boolean; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Color + * @description Color enhancement factor + * @default 1 */ - clip2?: components["schemas"]["CLIPField"] | null; + color?: number; + /** + * Contrast + * @description Contrast enhancement factor + * @default 1 + */ + contrast?: number; + /** + * Brightness + * @description Brightness enhancement factor + * @default 1 + */ + brightness?: number; + /** + * Sharpness + * @description Sharpness enhancement factor + * @default 1 + */ + sharpness?: number; /** * type - * @default metadata_to_sdlx_loras + * @default invokeai_img_enhance * @constant */ - type: "metadata_to_sdlx_loras"; + type: "invokeai_img_enhance"; }; /** - * Metadata To SDXL Model - * @description Extracts a SDXL Model value of a label from metadata + * Image Value Thresholds + * @description Clip image to pure black/white past specified thresholds. Originally created by @dwringer */ - MetadataToSDXLModelInvocation: { + InvokeImageValueThresholdsInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -30343,82 +15355,68 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default model - * @enum {string} - */ - label?: "* CUSTOM LABEL *" | "model"; - /** - * Custom Label - * @description Label for this metadata item - * @default null - */ - custom_label?: string | null; - /** - * @description The default SDXL Model to use if not found in the metadata + * @description The image from which to create a mask * @default null */ - default_value?: components["schemas"]["ModelIdentifierField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * type - * @default metadata_to_sdxl_model - * @constant + * Invert Output + * @description Make light areas dark and vice versa + * @default false */ - type: "metadata_to_sdxl_model"; - }; - /** - * MetadataToSDXLModelOutput - * @description String to SDXL main model output - */ - MetadataToSDXLModelOutput: { + invert_output?: boolean; /** - * Model - * @description Main model (UNet, VAE, CLIP) to load + * Renormalize Values + * @description Rescale remaining values from minimum to maximum + * @default false */ - model: components["schemas"]["ModelIdentifierField"]; + renormalize_values?: boolean; /** - * Name - * @description Model Name + * Lightness Only + * @description If true, only applies to image lightness (CIELa*b*) + * @default false */ - name: string; + lightness_only?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) + * Threshold Upper + * @description Threshold above which will be set to full value + * @default 0.5 */ - unet: components["schemas"]["UNetField"]; + threshold_upper?: number; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Threshold Lower + * @description Threshold below which will be set to minimum value + * @default 0.5 */ - clip: components["schemas"]["CLIPField"]; + threshold_lower?: number; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * type + * @default invokeai_img_val_thresholds + * @constant */ - clip2: components["schemas"]["CLIPField"]; + type: "invokeai_img_val_thresholds"; + }; + /** + * ItemIdsResult + * @description Response containing ordered item ids with metadata for optimistic updates. + */ + ItemIdsResult: { /** - * VAE - * @description VAE + * Item Ids + * @description Ordered list of item ids */ - vae: components["schemas"]["VAEField"]; - /** - * type - * @default metadata_to_sdxl_model_output - * @constant + item_ids: number[]; + /** + * Total Count + * @description Total number of queue items matching the query */ - type: "metadata_to_sdxl_model_output"; + total_count: number; }; /** - * Metadata To Scheduler - * @description Extracts a Scheduler value of a label from metadata + * IterateInvocation + * @description Iterates over a list of items */ - MetadataToSchedulerInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + IterateInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -30437,37 +15435,62 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default scheduler - * @enum {string} + * Collection + * @description The list of items to iterate over + * @default [] */ - label?: "* CUSTOM LABEL *" | "scheduler"; + collection?: unknown[]; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Index + * @description The index, will be provided on executed iterators + * @default 0 */ - custom_label?: string | null; + index?: number; /** - * Default Value - * @description The default scheduler to use if not found in the metadata - * @default euler - * @enum {string} + * type + * @default iterate + * @constant */ - default_value?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + type: "iterate"; + }; + /** + * IterateInvocationOutput + * @description Used to connect iteration outputs. Will be expanded to a specific output. + */ + IterateInvocationOutput: { + /** + * Collection Item + * @description The item being iterated over + */ + item: unknown; + /** + * Index + * @description The index of the item + */ + index: number; + /** + * Total + * @description The total number of items + */ + total: number; /** * type - * @default metadata_to_scheduler + * @default iterate_output * @constant */ - type: "metadata_to_scheduler"; + type: "iterate_output"; }; + JsonValue: unknown; /** - * Metadata To String Collection - * @description Extracts a string collection value of a label from metadata + * LaMa Infill + * @description Infills transparent areas of an image using the LaMa model */ - MetadataToStringCollectionInvocation: { + LaMaInfillInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; /** * @description Optional metadata to be saved with the image * @default null @@ -30491,41 +15514,22 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} - */ - label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt"; - /** - * Custom Label - * @description Label for this metadata item - * @default null - */ - custom_label?: string | null; - /** - * Default Value - * @description The default string collection to use if not found in the metadata + * @description The image to process * @default null */ - default_value?: string[] | null; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default metadata_to_string_collection + * @default infill_lama * @constant */ - type: "metadata_to_string_collection"; + type: "infill_lama"; }; /** - * Metadata To String - * @description Extracts a string value of a label from metadata + * Latents Collection Primitive + * @description A collection of latents tensor primitive values */ - MetadataToStringInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + LatentsCollectionInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -30544,41 +15548,57 @@ export type components = { */ use_cache?: boolean; /** - * Label - * @description Label for this metadata item - * @default * CUSTOM LABEL * - * @enum {string} + * Collection + * @description The collection of latents tensors + * @default null */ - label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt"; + collection?: components["schemas"]["LatentsField"][] | null; /** - * Custom Label - * @description Label for this metadata item - * @default null + * type + * @default latents_collection + * @constant */ - custom_label?: string | null; + type: "latents_collection"; + }; + /** + * LatentsCollectionOutput + * @description Base class for nodes that output a collection of latents tensors + */ + LatentsCollectionOutput: { /** - * Default Value - * @description The default string to use if not found in the metadata - * @default null + * Collection + * @description Latents tensor */ - default_value?: string | null; + collection: components["schemas"]["LatentsField"][]; /** * type - * @default metadata_to_string + * @default latents_collection_output * @constant */ - type: "metadata_to_string"; + type: "latents_collection_output"; }; /** - * Metadata To T2I-Adapters - * @description Extracts a T2I-Adapters value of a label from metadata + * LatentsField + * @description A latents tensor primitive field */ - MetadataToT2IAdaptersInvocation: { + LatentsField: { /** - * @description Optional metadata to be saved with the image + * Latents Name + * @description The name of the latents + */ + latents_name: string; + /** + * Seed + * @description Seed used to generate this latents * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + seed?: number | null; + }; + /** + * Latents Primitive + * @description A latents tensor primitive value + */ + LatentsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -30597,76 +15617,82 @@ export type components = { */ use_cache?: boolean; /** - * T2I-Adapter - * @description IP-Adapter to apply + * @description The latents tensor * @default null */ - t2i_adapter_list?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; + latents?: components["schemas"]["LatentsField"] | null; /** * type - * @default metadata_to_t2i_adapters + * @default latents * @constant */ - type: "metadata_to_t2i_adapters"; + type: "latents"; }; /** - * Metadata To VAE - * @description Extracts a VAE value of a label from metadata + * LatentsMetaOutput + * @description Latents + metadata */ - MetadataToVAEInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + LatentsMetaOutput: { + /** @description Metadata Dict */ + metadata: components["schemas"]["MetadataField"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * type + * @default latents_meta_output + * @constant */ - is_intermediate?: boolean; + type: "latents_meta_output"; + /** @description Latents tensor */ + latents: components["schemas"]["LatentsField"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Width + * @description Width of output (px) */ - use_cache?: boolean; + width: number; /** - * Label - * @description Label for this metadata item - * @default vae - * @enum {string} + * Height + * @description Height of output (px) */ - label?: "* CUSTOM LABEL *" | "vae"; + height: number; + }; + /** + * LatentsOutput + * @description Base class for nodes that output a single latents tensor + */ + LatentsOutput: { + /** @description Latents tensor */ + latents: components["schemas"]["LatentsField"]; /** - * Custom Label - * @description Label for this metadata item - * @default null + * Width + * @description Width of output (px) */ - custom_label?: string | null; + width: number; /** - * @description The default VAE to use if not found in the metadata - * @default null + * Height + * @description Height of output (px) */ - default_value?: components["schemas"]["VAEField"] | null; + height: number; /** * type - * @default metadata_to_vae + * @default latents_output * @constant */ - type: "metadata_to_vae"; + type: "latents_output"; }; /** - * Minimum Overlap XYImage Tile Generator - * @description Cuts up an image into overlapping tiles and outputs a string representation of the tiles to use, taking the - * input overlap as a minimum + * Latents to Image - SD1.5, SDXL + * @description Generates an image from latents. */ - MinimumOverlapXYTileGenerator: { + LatentsToImageInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -30685,81 +15711,55 @@ export type components = { */ use_cache?: boolean; /** - * @description The input image + * @description Latents tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * Tile Width - * @description x resolution of generation tile (must be a multiple of 8) - * @default 576 + * @description VAE + * @default null */ - tile_width?: number; + vae?: components["schemas"]["VAEField"] | null; /** - * Tile Height - * @description y resolution of generation tile (must be a multiple of 8) - * @default 576 + * Tiled + * @description Processing using overlapping tiles (reduce memory consumption) + * @default false */ - tile_height?: number; + tiled?: boolean; /** - * Min Overlap - * @description minimum tile overlap size (must be a multiple of 8) - * @default 128 + * Tile Size + * @description The tile size for VAE tiling in pixels (image space). If set to 0, the default tile size for the model will be used. Larger tile sizes generally produce better results at the cost of higher memory usage. + * @default 0 */ - min_overlap?: number; + tile_size?: number; /** - * Round To 8 - * @description Round outputs down to the nearest 8 (for pulling from a large noise field) + * Fp32 + * @description Whether or not to use full float32 precision * @default false */ - round_to_8?: boolean; + fp32?: boolean; /** * type - * @default minimum_overlap_xy_tile_generator + * @default l2i * @constant */ - type: "minimum_overlap_xy_tile_generator"; + type: "l2i"; }; /** - * ModelFormat - * @description Storage format of model. - * @enum {string} + * Lineart Anime Edge Detection + * @description Geneartes an edge map using the Lineart model. */ - ModelFormat: "omi" | "diffusers" | "checkpoint" | "lycoris" | "onnx" | "olive" | "embedding_file" | "embedding_folder" | "invokeai" | "t5_encoder" | "qwen3_encoder" | "bnb_quantized_int8b" | "bnb_quantized_nf4b" | "gguf_quantized" | "unknown"; - /** ModelIdentifierField */ - ModelIdentifierField: { - /** - * Key - * @description The model's unique key - */ - key: string; - /** - * Hash - * @description The model's BLAKE3 hash - */ - hash: string; + LineartAnimeEdgeDetectionInvocation: { /** - * Name - * @description The model's name + * @description The board to save the image to + * @default null */ - name: string; - /** @description The model's base model type */ - base: components["schemas"]["BaseModelType"]; - /** @description The model's type */ - type: components["schemas"]["ModelType"]; + board?: components["schemas"]["BoardField"] | null; /** - * @description The submodel to load, if this is a main model + * @description Optional metadata to be saved with the image * @default null */ - submodel_type?: components["schemas"]["SubModelType"] | null; - }; - /** - * Any Model - * @description Selects any model, outputting it its identifier. Be careful with this one! The identifier will be accepted as - * input for any model, even if the model types don't match. If you connect this to a mismatched input, you'll get an - * error. - */ - ModelIdentifierInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -30778,402 +15778,367 @@ export type components = { */ use_cache?: boolean; /** - * Model - * @description The model to select + * @description The image to process * @default null */ - model?: components["schemas"]["ModelIdentifierField"] | null; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default model_identifier + * @default lineart_anime_edge_detection * @constant */ - type: "model_identifier"; + type: "lineart_anime_edge_detection"; }; /** - * ModelIdentifierOutput - * @description Model identifier output + * Lineart Edge Detection + * @description Generates an edge map using the Lineart model. */ - ModelIdentifierOutput: { - /** - * Model - * @description Model identifier - */ - model: components["schemas"]["ModelIdentifierField"]; + LineartEdgeDetectionInvocation: { /** - * type - * @default model_identifier_output - * @constant + * @description The board to save the image to + * @default null */ - type: "model_identifier_output"; - }; - /** - * ModelInstallCancelledEvent - * @description Event model for model_install_cancelled - */ - ModelInstallCancelledEvent: { + board?: components["schemas"]["BoardField"] | null; /** - * Timestamp - * @description The timestamp of the event + * @description Optional metadata to be saved with the image + * @default null */ - timestamp: number; + metadata?: components["schemas"]["MetadataField"] | null; /** * Id - * @description The ID of the install job - */ - id: number; - /** - * Source - * @description Source of the model; local path, repo_id or url - */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; - }; - /** - * ModelInstallCompleteEvent - * @description Event model for model_install_complete - */ - ModelInstallCompleteEvent: { - /** - * Timestamp - * @description The timestamp of the event + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - timestamp: number; + id: string; /** - * Id - * @description The ID of the install job + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - id: number; + is_intermediate?: boolean; /** - * Source - * @description Source of the model; local path, repo_id or url + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + use_cache?: boolean; /** - * Key - * @description Model config record key + * @description The image to process + * @default null */ - key: string; + image?: components["schemas"]["ImageField"] | null; /** - * Total Bytes - * @description Size of the model (may be None for installation of a local path) + * Coarse + * @description Whether to use coarse mode + * @default false */ - total_bytes: number | null; + coarse?: boolean; /** - * Config - * @description The installed model's config + * type + * @default lineart_edge_detection + * @constant */ - config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; + type: "lineart_edge_detection"; }; /** - * ModelInstallDownloadProgressEvent - * @description Event model for model_install_download_progress + * LLaVA OneVision VLLM + * @description Run a LLaVA OneVision VLLM model. */ - ModelInstallDownloadProgressEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; + LlavaOnevisionVllmInvocation: { /** * Id - * @description The ID of the install job - */ - id: number; - /** - * Source - * @description Source of the model; local path, repo_id or url - */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; - /** - * Local Path - * @description Where model is downloading to - */ - local_path: string; - /** - * Bytes - * @description Number of bytes downloaded so far - */ - bytes: number; - /** - * Total Bytes - * @description Total size of download, including all files - */ - total_bytes: number; - /** - * Parts - * @description Progress of downloading URLs that comprise the model, if any - */ - parts: { - [key: string]: number | string; - }[]; - }; - /** - * ModelInstallDownloadStartedEvent - * @description Event model for model_install_download_started - */ - ModelInstallDownloadStartedEvent: { - /** - * Timestamp - * @description The timestamp of the event + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - timestamp: number; + id: string; /** - * Id - * @description The ID of the install job + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - id: number; + is_intermediate?: boolean; /** - * Source - * @description Source of the model; local path, repo_id or url + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + use_cache?: boolean; /** - * Local Path - * @description Where model is downloading to + * Images + * @description Input image. + * @default null */ - local_path: string; + images?: (components["schemas"]["ImageField"][] | components["schemas"]["ImageField"]) | null; /** - * Bytes - * @description Number of bytes downloaded so far + * Prompt + * @description Input text prompt. + * @default */ - bytes: number; + prompt?: string; /** - * Total Bytes - * @description Total size of download, including all files + * LLaVA Model Type + * @description The VLLM model to use + * @default null */ - total_bytes: number; - /** - * Parts - * @description Progress of downloading URLs that comprise the model, if any + vllm_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default llava_onevision_vllm + * @constant */ - parts: { - [key: string]: number | string; - }[]; + type: "llava_onevision_vllm"; }; /** - * ModelInstallDownloadsCompleteEvent - * @description Emitted once when an install job becomes active. + * LlavaOnevision_Diffusers_Config + * @description Model config for Llava Onevision models. */ - ModelInstallDownloadsCompleteEvent: { + LlavaOnevision_Diffusers_Config: { /** - * Timestamp - * @description The timestamp of the event + * Key + * @description A unique key for this model. */ - timestamp: number; + key: string; /** - * Id - * @description The ID of the install job + * Hash + * @description The hash of the model file(s). */ - id: number; + hash: string; /** - * Source - * @description Source of the model; local path, repo_id or url + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; - }; - /** - * ModelInstallErrorEvent - * @description Event model for model_install_error - */ - ModelInstallErrorEvent: { + path: string; /** - * Timestamp - * @description The timestamp of the event + * File Size + * @description The size of the model in bytes. */ - timestamp: number; + file_size: number; /** - * Id - * @description The ID of the install job + * Name + * @description Name of the model. */ - id: number; + name: string; /** - * Source - * @description Source of the model; local path, repo_id or url + * Description + * @description Model description */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + description: string | null; /** - * Error Type - * @description The name of the exception + * Source + * @description The original source of the model (path, URL or repo_id). */ - error_type: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Error - * @description A text description of the exception + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - error: string; - }; - /** - * ModelInstallJob - * @description Object that tracks the current status of an install request. - */ - ModelInstallJob: { + source_api_response: string | null; /** - * Id - * @description Unique ID for this job + * Cover Image + * @description Url for image to preview model */ - id: number; + cover_image: string | null; /** - * @description Current status of install process - * @default waiting + * Format + * @default diffusers + * @constant */ - status?: components["schemas"]["InstallStatus"]; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Error Reason - * @description Information about why the job failed + * Type + * @default llava_onevision + * @constant */ - error_reason?: string | null; - /** @description Configuration information (e.g. 'description') to apply to model. */ - config_in?: components["schemas"]["ModelRecordChanges"]; + type: "llava_onevision"; /** - * Config Out - * @description After successful installation, this will hold the configuration object. + * Base + * @default any + * @constant */ - config_out?: (components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]) | null; + base: "any"; /** - * Inplace - * @description Leave model in its current location; otherwise install under models directory - * @default false + * Cpu Only + * @description Whether this model should run on CPU only */ - inplace?: boolean; + cpu_only: boolean | null; + }; + /** + * Apply LoRA Collection - SD1.5 + * @description Applies a collection of LoRAs to the provided UNet and CLIP models. + */ + LoRACollectionLoader: { /** - * Source - * @description Source (URL, repo_id, or local path) of model + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + id: string; /** - * Local Path - * Format: path - * @description Path to locally-downloaded model; may be the same as the source + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - local_path: string; + is_intermediate?: boolean; /** - * Bytes - * @description For a remote model, the number of bytes downloaded so far (may not be available) - * @default 0 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - bytes?: number; + use_cache?: boolean; /** - * Total Bytes - * @description Total size of the model to be installed - * @default 0 + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. + * @default null */ - total_bytes?: number; + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * Source Metadata - * @description Metadata provided by the model source + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - source_metadata?: (components["schemas"]["BaseMetadata"] | components["schemas"]["HuggingFaceMetadata"]) | null; + unet?: components["schemas"]["UNetField"] | null; /** - * Download Parts - * @description Download jobs contributing to this install + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - download_parts?: components["schemas"]["DownloadJob"][]; + clip?: components["schemas"]["CLIPField"] | null; /** - * Error - * @description On an error condition, this field will contain the text of the exception + * type + * @default lora_collection_loader + * @constant */ - error?: string | null; + type: "lora_collection_loader"; + }; + /** LoRAField */ + LoRAField: { + /** @description Info to load lora model */ + lora: components["schemas"]["ModelIdentifierField"]; /** - * Error Traceback - * @description On an error condition, this field will contain the exception traceback + * Weight + * @description Weight to apply to lora model */ - error_traceback?: string | null; + weight: number; }; /** - * ModelInstallStartedEvent - * @description Event model for model_install_started + * Apply LoRA - SD1.5 + * @description Apply selected lora to unet and text_encoder. */ - ModelInstallStartedEvent: { + LoRALoaderInvocation: { /** - * Timestamp - * @description The timestamp of the event + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - timestamp: number; + id: string; /** - * Id - * @description The ID of the install job + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - id: number; + is_intermediate?: boolean; /** - * Source - * @description Source of the model; local path, repo_id or url + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; - }; - /** - * ModelLoadCompleteEvent - * @description Event model for model_load_complete - */ - ModelLoadCompleteEvent: { + use_cache?: boolean; /** - * Timestamp - * @description The timestamp of the event + * LoRA + * @description LoRA model to load + * @default null */ - timestamp: number; + lora?: components["schemas"]["ModelIdentifierField"] | null; /** - * Config - * @description The model's config + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 */ - config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; + weight?: number; /** - * @description The submodel type, if any + * UNet + * @description UNet (scheduler, LoRAs) * @default null */ - submodel_type: components["schemas"]["SubModelType"] | null; + unet?: components["schemas"]["UNetField"] | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * type + * @default lora_loader + * @constant + */ + type: "lora_loader"; }; /** - * ModelLoadStartedEvent - * @description Event model for model_load_started + * LoRALoaderOutput + * @description Model loader output */ - ModelLoadStartedEvent: { + LoRALoaderOutput: { /** - * Timestamp - * @description The timestamp of the event + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - timestamp: number; + unet: components["schemas"]["UNetField"] | null; /** - * Config - * @description The model's config + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; + clip: components["schemas"]["CLIPField"] | null; /** - * @description The submodel type, if any - * @default null + * type + * @default lora_loader_output + * @constant */ - submodel_type: components["schemas"]["SubModelType"] | null; + type: "lora_loader_output"; }; /** - * ModelLoaderOutput - * @description Model loader output + * LoRAMetadataField + * @description LoRA Metadata Field */ - ModelLoaderOutput: { + LoRAMetadataField: { + /** @description LoRA model to load */ + model: components["schemas"]["ModelIdentifierField"]; /** - * VAE - * @description VAE + * Weight + * @description The weight at which the LoRA is applied to each model */ - vae: components["schemas"]["VAEField"]; + weight: number; + }; + /** + * LoRARecallParameter + * @description LoRA configuration for recall + */ + LoRARecallParameter: { /** - * type - * @default model_loader_output - * @constant + * Model Name + * @description The name of the LoRA model */ - type: "model_loader_output"; + model_name: string; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Weight + * @description The weight for the LoRA + * @default 0.75 */ - clip: components["schemas"]["CLIPField"]; + weight?: number; /** - * UNet - * @description UNet (scheduler, LoRAs) + * Is Enabled + * @description Whether the LoRA is enabled + * @default true */ - unet: components["schemas"]["UNetField"]; + is_enabled?: boolean; }; /** - * Model Name Grabber - * @description Outputs the Model's name as a string. Use the Model Identifier node as input. + * Select LoRA + * @description Selects a LoRA model and weight. */ - ModelNameGrabberInvocation: { + LoRASelectorInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -31192,1183 +16157,1236 @@ export type components = { */ use_cache?: boolean; /** - * Model - * @description Get this input from a Model Identifier node. + * LoRA + * @description LoRA model to load * @default null */ - model?: components["schemas"]["ModelIdentifierField"] | null; + lora?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 + */ + weight?: number; /** * type - * @default model_name_grabber_invocation + * @default lora_selector * @constant */ - type: "model_name_grabber_invocation"; + type: "lora_selector"; }; /** - * ModelNameGrabberOutput - * @description Model Name Grabber output + * LoRASelectorOutput + * @description Model loader output */ - ModelNameGrabberOutput: { + LoRASelectorOutput: { /** - * Name - * @description Name of the Model + * LoRA + * @description LoRA model and weight */ - name: string; + lora: components["schemas"]["LoRAField"]; /** * type - * @default model_name_grabber_output + * @default lora_selector_output * @constant */ - type: "model_name_grabber_output"; + type: "lora_selector_output"; }; - /** - * Model Name to Model - * @description Converts a model identified by its string name to a Model - */ - ModelNameToModelInvocation: { + /** LoRA_Diffusers_FLUX_Config */ + LoRA_Diffusers_FLUX_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Model Name - * @description Exact model name (must match exactly one model) - * @default null + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. */ - model_name?: string | null; + name: string; /** - * type - * @default model_name_to_model + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** + * Base + * @default flux * @constant */ - type: "model_name_to_model"; + base: "flux"; }; /** - * ModelRecordChanges - * @description A set of changes to apply to a model. + * LoRA_Diffusers_Flux2_Config + * @description Model config for FLUX.2 (Klein) LoRA models in Diffusers format. */ - ModelRecordChanges: { + LoRA_Diffusers_Flux2_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + */ + path: string; + /** + * File Size + * @description The size of the model in bytes. + */ + file_size: number; + /** + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; /** * Source - * @description original source of the model + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora + * @constant + */ + type: "lora"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** + * Base + * @default flux2 + * @constant + */ + base: "flux2"; + variant: components["schemas"]["Flux2VariantType"] | null; + }; + /** LoRA_Diffusers_SD1_Config */ + LoRA_Diffusers_SD1_Config: { + /** + * Key + * @description A unique key for this model. + */ + key: string; + /** + * Hash + * @description The hash of the model file(s). + */ + hash: string; + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - source?: string | null; - /** @description type of model source */ - source_type?: components["schemas"]["ModelSourceType"] | null; + path: string; /** - * Source Api Response - * @description metadata from remote source + * File Size + * @description The size of the model in bytes. */ - source_api_response?: string | null; + file_size: number; /** * Name * @description Name of the model. */ - name?: string | null; - /** - * Path - * @description Path to the model. - */ - path?: string | null; + name: string; /** * Description * @description Model description */ - description?: string | null; - /** @description The base model. */ - base?: components["schemas"]["BaseModelType"] | null; - /** @description Type of model */ - type?: components["schemas"]["ModelType"] | null; + description: string | null; /** - * Key - * @description Database ID for this model + * Source + * @description The original source of the model (path, URL or repo_id). */ - key?: string | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Hash - * @description hash of model file + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - hash?: string | null; + source_api_response: string | null; /** - * File Size - * @description Size of model file + * Cover Image + * @description Url for image to preview model */ - file_size?: number | null; + cover_image: string | null; /** - * Format - * @description format of model file + * Type + * @default lora + * @constant */ - format?: string | null; + type: "lora"; /** * Trigger Phrases * @description Set of trigger phrases for this model */ - trigger_phrases?: string[] | null; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * Default Settings - * @description Default settings for this model + * Format + * @default diffusers + * @constant */ - default_settings?: components["schemas"]["MainModelDefaultSettings"] | components["schemas"]["LoraModelDefaultSettings"] | components["schemas"]["ControlAdapterDefaultSettings"] | null; + format: "diffusers"; /** - * Cpu Only - * @description Whether this model should run on CPU only + * Base + * @default sd-1 + * @constant */ - cpu_only?: boolean | null; + base: "sd-1"; + }; + /** LoRA_Diffusers_SD2_Config */ + LoRA_Diffusers_SD2_Config: { /** - * Variant - * @description The variant of the model. + * Key + * @description A unique key for this model. */ - variant?: components["schemas"]["ModelVariantType"] | components["schemas"]["ClipVariantType"] | components["schemas"]["FluxVariantType"] | components["schemas"]["Flux2VariantType"] | components["schemas"]["ZImageVariantType"] | components["schemas"]["Qwen3VariantType"] | null; - /** @description The prediction type of the model. */ - prediction_type?: components["schemas"]["SchedulerPredictionType"] | null; + key: string; /** - * Upcast Attention - * @description Whether to upcast attention. + * Hash + * @description The hash of the model file(s). */ - upcast_attention?: boolean | null; + hash: string; /** - * Config Path - * @description Path to config file for model + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - config_path?: string | null; - }; - /** - * ModelRecordOrderBy - * @description The order in which to return model summaries. - * @enum {string} - */ - ModelRecordOrderBy: "default" | "type" | "base" | "name" | "format" | "size" | "created_at" | "updated_at" | "path"; - /** ModelRelationshipBatchRequest */ - ModelRelationshipBatchRequest: { + path: string; /** - * Model Keys - * @description List of model keys to fetch related models for + * File Size + * @description The size of the model in bytes. */ - model_keys: string[]; - }; - /** ModelRelationshipCreateRequest */ - ModelRelationshipCreateRequest: { + file_size: number; /** - * Model Key 1 - * @description The key of the first model in the relationship + * Name + * @description Name of the model. */ - model_key_1: string; + name: string; /** - * Model Key 2 - * @description The key of the second model in the relationship + * Description + * @description Model description */ - model_key_2: string; - }; - /** - * ModelRepoVariant - * @description Various hugging face variants on the diffusers format. - * @enum {string} - */ - ModelRepoVariant: "" | "fp16" | "fp32" | "onnx" | "openvino" | "flax"; - /** - * ModelSourceType - * @description Model source type. - * @enum {string} - */ - ModelSourceType: "path" | "url" | "hf_repo_id"; - /** - * Model To String - * @description Converts an Model to a JSONString - */ - ModelToStringInvocation: { + description: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source + * @description The original source of the model (path, URL or repo_id). */ - id: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - is_intermediate?: boolean; + source_api_response: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Cover Image + * @description Url for image to preview model */ - use_cache?: boolean; + cover_image: string | null; /** - * Model - * @description The model to select - * @default null + * Type + * @default lora + * @constant */ - model?: components["schemas"]["ModelIdentifierField"] | null; + type: "lora"; /** - * type - * @default model_to_string + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers * @constant */ - type: "model_to_string"; + format: "diffusers"; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; }; - /** - * Model Toggle - * @description Allows boolean selection between two separate ModelIdentifier inputs - */ - ModelToggleInvocation: { + /** LoRA_Diffusers_SDXL_Config */ + LoRA_Diffusers_SDXL_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Use Second - * @description Use 2nd Input - * @default null + * File Size + * @description The size of the model in bytes. */ - use_second?: boolean | null; + file_size: number; /** - * @description First Model Input - * @default null + * Name + * @description Name of the model. */ - model1?: components["schemas"]["ModelIdentifierField"] | null; + name: string; /** - * @description First Model Input - * @default null + * Description + * @description Model description */ - model2?: components["schemas"]["ModelIdentifierField"] | null; + description: string | null; /** - * type - * @default model_toggle + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora * @constant */ - type: "model_toggle"; - }; - /** - * ModelToggleOutput - * @description Model Toggle output - */ - ModelToggleOutput: { + type: "lora"; /** - * Model - * @description Model identifier + * Trigger Phrases + * @description Set of trigger phrases for this model */ - model: components["schemas"]["ModelIdentifierField"]; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; /** - * type - * @default model_toggle_output + * Base + * @default sdxl * @constant */ - type: "model_toggle_output"; - }; - /** - * ModelType - * @description Model type. - * @enum {string} - */ - ModelType: "onnx" | "main" | "vae" | "lora" | "control_lora" | "controlnet" | "embedding" | "ip_adapter" | "clip_vision" | "clip_embed" | "t2i_adapter" | "t5_encoder" | "qwen3_encoder" | "spandrel_image_to_image" | "siglip" | "flux_redux" | "llava_onevision" | "unknown"; - /** - * ModelVariantType - * @description Variant type. - * @enum {string} - */ - ModelVariantType: "normal" | "inpaint" | "depth"; - /** - * ModelsList - * @description Return list of configs. - */ - ModelsList: { - /** Models */ - models: (components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"])[]; + base: "sdxl"; }; /** - * MonochromeFilmGrain - * @description Adds monochrome film grain to an image + * LoRA_Diffusers_ZImage_Config + * @description Model config for Z-Image LoRA models in Diffusers format. */ - MonochromeFilmGrainInvocation: { + LoRA_Diffusers_ZImage_Config: { /** - * @description The board to save the image to - * @default null + * Key + * @description A unique key for this model. */ - board?: components["schemas"]["BoardField"] | null; + key: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Hash + * @description The hash of the model file(s). */ - metadata?: components["schemas"]["MetadataField"] | null; + hash: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - id: string; + path: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * File Size + * @description The size of the model in bytes. */ - is_intermediate?: boolean; + file_size: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Name + * @description Name of the model. */ - use_cache?: boolean; + name: string; /** - * @description The image to add film grain to - * @default null + * Description + * @description Model description */ - image?: components["schemas"]["ImageField"] | null; + description: string | null; /** - * Amount 1 - * @description Amount of the first noise layer - * @default 100 + * Source + * @description The original source of the model (path, URL or repo_id). */ - amount_1?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Amount 2 - * @description Amount of the second noise layer - * @default 50 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - amount_2?: number; + source_api_response: string | null; /** - * Seed 1 - * @description The first seed to use (omit for random) - * @default null + * Cover Image + * @description Url for image to preview model */ - seed_1?: number | null; + cover_image: string | null; /** - * Seed 2 - * @description The second seed to use (omit for random) - * @default null + * Type + * @default lora + * @constant */ - seed_2?: number | null; + type: "lora"; /** - * Blur 1 - * @description The strength of the first noise blur - * @default 0.5 + * Trigger Phrases + * @description Set of trigger phrases for this model */ - blur_1?: number; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * Blur 2 - * @description The strength of the second noise blur - * @default 0.5 + * Format + * @default diffusers + * @constant */ - blur_2?: number; + format: "diffusers"; /** - * type - * @default monochrome_film_grain + * Base + * @default z-image * @constant */ - type: "monochrome_film_grain"; + base: "z-image"; + variant: components["schemas"]["ZImageVariantType"] | null; }; - /** - * Multiply Integers - * @description Multiplies two numbers - */ - MultiplyInvocation: { + /** LoRA_LyCORIS_FLUX_Config */ + LoRA_LyCORIS_FLUX_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * A - * @description The first number - * @default 0 + * File Size + * @description The size of the model in bytes. */ - a?: number; + file_size: number; /** - * B - * @description The second number - * @default 0 + * Name + * @description Name of the model. */ - b?: number; + name: string; /** - * type - * @default mul - * @constant + * Description + * @description Model description */ - type: "mul"; - }; - /** - * Nightmare Promptgen - * @description makes new friends - */ - NightmareInvocation: { + description: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source + * @description The original source of the model (path, URL or repo_id). */ - id: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - is_intermediate?: boolean; + source_api_response: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Cover Image + * @description Url for image to preview model */ - use_cache?: boolean; + cover_image: string | null; /** - * Prompt - * @description starting point for the generated prompt - * @default + * Type + * @default lora + * @constant */ - prompt?: string; + type: "lora"; /** - * Max New Tokens - * @description the maximum allowed amount of new tokens to generate - * @default 300 + * Trigger Phrases + * @description Set of trigger phrases for this model */ - max_new_tokens?: number; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * Min New Tokens - * @description the minimum new tokens - NOTE, this can increase generation time - * @default 30 + * Format + * @default lycoris + * @constant */ - min_new_tokens?: number; + format: "lycoris"; /** - * Max Time - * @description Overrules min tokens; the max amount of time allowed to generate - * @default 10 + * Base + * @default flux + * @constant */ - max_time?: number; + base: "flux"; + }; + /** + * LoRA_LyCORIS_Flux2_Config + * @description Model config for FLUX.2 (Klein) LoRA models in LyCORIS format. + */ + LoRA_LyCORIS_Flux2_Config: { /** - * Temp - * @description Temperature - * @default 1.8 + * Key + * @description A unique key for this model. */ - temp?: number; + key: string; /** - * Typical P - * @description Lower than 1.0 seems crazier, higher is more consistent. - * @default 1 + * Hash + * @description The hash of the model file(s). */ - typical_p?: number; + hash: string; /** - * Top P - * @description Top P sampling - * @default 0.9 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - top_p?: number; + path: string; /** - * Top K - * @description Top K sampling - * @default 20 + * File Size + * @description The size of the model in bytes. */ - top_k?: number; + file_size: number; /** - * Repetition Penalty - * @description Higher than 1.0 will try to prevent repetition. - * @default 1 + * Name + * @description Name of the model. */ - repetition_penalty?: number; + name: string; /** - * Include Starter - * @description Include your prompt starter with the output or not. - * @default true + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). */ - include_starter?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Repo Id - * @default cactusfriend/nightmare-promptgen-3 - * @enum {string} + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - repo_id?: "cactusfriend/nightmare-promptgen-3" | "cactusfriend/nightmare-promptgen-XL" | "cactusfriend/nightmare-invokeai-prompts"; + source_api_response: string | null; /** - * type - * @default nightmare_promptgen + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora * @constant */ - type: "nightmare_promptgen"; - }; - /** - * NightmareOutput - * @description Nightmare prompt string output - */ - NightmareOutput: { + type: "lora"; /** - * Prompt - * @description The generated nightmare prompt string + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; + /** + * Format + * @default lycoris + * @constant */ - prompt: string; + format: "lycoris"; /** - * type - * @default nightmare_str_output + * Base + * @default flux2 * @constant */ - type: "nightmare_str_output"; + base: "flux2"; + variant: components["schemas"]["Flux2VariantType"] | null; }; - /** NodeFieldValue */ - NodeFieldValue: { + /** LoRA_LyCORIS_SD1_Config */ + LoRA_LyCORIS_SD1_Config: { /** - * Node Path - * @description The node into which this batch data item will be substituted. + * Key + * @description A unique key for this model. */ - node_path: string; + key: string; /** - * Field Name - * @description The field into which this batch data item will be substituted. + * Hash + * @description The hash of the model file(s). */ - field_name: string; + hash: string; /** - * Value - * @description The value to substitute into the node/field. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - value: string | number | components["schemas"]["ImageField"]; - }; - /** - * Add Noise (Flux) - * @description Add noise to a flux latents tensor using the appropriate ratio given the denoising schedule timestep. - * - * Calculates the correct initial timestep noising amount and applies it to the given latent tensor using simple addition according to the specified ratio. - */ - NoiseAddFluxInvocation: { + path: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * File Size + * @description The size of the model in bytes. */ - id: string; + file_size: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Name + * @description Name of the model. */ - is_intermediate?: boolean; + name: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Description + * @description Model description */ - use_cache?: boolean; + description: string | null; /** - * @description Latents tensor - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - latents_in?: components["schemas"]["LatentsField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * @description The noise to be added - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - noise_in?: components["schemas"]["LatentsField"] | null; + source_api_response: string | null; /** - * Num Steps - * @description Number of diffusion steps - * @default null + * Cover Image + * @description Url for image to preview model */ - num_steps?: number | null; + cover_image: string | null; /** - * Denoising Start - * @description Starting point for denoising (0.0 to 1.0) - * @default null + * Type + * @default lora + * @constant */ - denoising_start?: number | null; + type: "lora"; /** - * Is Schnell - * @description Boolean flag indicating if this is a FLUX Schnell model - * @default null + * Trigger Phrases + * @description Set of trigger phrases for this model */ - is_schnell?: boolean | null; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * type - * @default noise_add_flux + * Format + * @default lycoris * @constant */ - type: "noise_add_flux"; - }; - /** - * 2D Noise Image - * @description Creates an image of 2D Noise approximating the desired characteristics. - * - * Creates an image of 2D Noise approximating the desired characteristics, using various combinations of gaussian blur and arithmetic operations to perform low pass and high pass filtering of 2-dimensional spatial frequencies of each channel to create Red, Blue, or Green "colored noise". - */ - NoiseImage2DInvocation: { + format: "lycoris"; /** - * @description The board to save the image to - * @default null + * Base + * @default sd-1 + * @constant */ - board?: components["schemas"]["BoardField"] | null; + base: "sd-1"; + }; + /** LoRA_LyCORIS_SD2_Config */ + LoRA_LyCORIS_SD2_Config: { /** - * @description Optional metadata to be saved with the image - * @default null + * Key + * @description A unique key for this model. */ - metadata?: components["schemas"]["MetadataField"] | null; + key: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Hash + * @description The hash of the model file(s). */ - id: string; + hash: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - is_intermediate?: boolean; + path: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * File Size + * @description The size of the model in bytes. */ - use_cache?: boolean; + file_size: number; /** - * Noise Type - * @description Desired noise spectral characteristics - * @default White - * @enum {string} + * Name + * @description Name of the model. */ - noise_type?: "White" | "Red" | "Blue" | "Green"; + name: string; /** - * Width - * @description Desired image width - * @default 512 + * Description + * @description Model description */ - width?: number; + description: string | null; /** - * Height - * @description Desired image height - * @default 512 + * Source + * @description The original source of the model (path, URL or repo_id). */ - height?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Seed - * @description Seed for noise generation - * @default 0 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - seed?: number; + source_api_response: string | null; /** - * Iterations - * @description Noise approx. iterations - * @default 15 + * Cover Image + * @description Url for image to preview model */ - iterations?: number; + cover_image: string | null; /** - * Blur Threshold - * @description Threshold used in computing noise (lower is better/slower) - * @default 0.2 + * Type + * @default lora + * @constant */ - blur_threshold?: number; + type: "lora"; /** - * Sigma Red - * @description Sigma for strong gaussian blur LPF for red/green - * @default 3 + * Trigger Phrases + * @description Set of trigger phrases for this model */ - sigma_red?: number; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * Sigma Blue - * @description Sigma for weak gaussian blur HPF for blue/green - * @default 1 + * Format + * @default lycoris + * @constant */ - sigma_blue?: number; + format: "lycoris"; /** - * type - * @default noiseimg_2d + * Base + * @default sd-2 * @constant */ - type: "noiseimg_2d"; + base: "sd-2"; }; - /** - * Create Latent Noise - * @description Generates latent noise. - */ - NoiseInvocation: { + /** LoRA_LyCORIS_SDXL_Config */ + LoRA_LyCORIS_SDXL_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Seed - * @description Seed for random number generation - * @default 0 + * File Size + * @description The size of the model in bytes. */ - seed?: number; + file_size: number; /** - * Width - * @description Width of output (px) - * @default 512 + * Name + * @description Name of the model. */ - width?: number; + name: string; /** - * Height - * @description Height of output (px) - * @default 512 + * Description + * @description Model description */ - height?: number; + description: string | null; /** - * Use Cpu - * @description Use CPU for noise generation (for reproducible results across platforms) - * @default true + * Source + * @description The original source of the model (path, URL or repo_id). */ - use_cpu?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * type - * @default noise + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default lora * @constant */ - type: "noise"; - }; - /** - * NoiseOutput - * @description Invocation noise output - */ - NoiseOutput: { - /** @description Noise tensor */ - noise: components["schemas"]["LatentsField"]; + type: "lora"; /** - * Width - * @description Width of output (px) + * Trigger Phrases + * @description Set of trigger phrases for this model */ - width: number; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * Height - * @description Height of output (px) + * Format + * @default lycoris + * @constant */ - height: number; + format: "lycoris"; /** - * type - * @default noise_output + * Base + * @default sdxl * @constant */ - type: "noise_output"; + base: "sdxl"; }; /** - * Noise (Spectral characteristics) - * @description Creates a latents tensor of 2D noise channels approximating the desired characteristics. - * - * This operates like 2D Noise Image but outputs latent tensors, 4-channel or 16-channel. + * LoRA_LyCORIS_ZImage_Config + * @description Model config for Z-Image LoRA models in LyCORIS format. */ - NoiseSpectralInvocation: { + LoRA_LyCORIS_ZImage_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Noise Type - * @description Desired noise spectral characteristics - * @default White - * @enum {string} + * File Size + * @description The size of the model in bytes. */ - noise_type?: "White" | "Red" | "Blue" | "Green"; + file_size: number; /** - * Width - * @description Desired image width - * @default 512 + * Name + * @description Name of the model. */ - width?: number; + name: string; /** - * Height - * @description Desired image height - * @default 512 + * Description + * @description Model description */ - height?: number; + description: string | null; /** - * Flux16 - * @description If false, 4-channel (SD/SDXL); if true, 16-channel (Flux) - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - flux16?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Seed - * @description Seed for noise generation - * @default 0 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - seed?: number; + source_api_response: string | null; /** - * Iterations - * @description Noise approx. iterations - * @default 15 + * Cover Image + * @description Url for image to preview model */ - iterations?: number; + cover_image: string | null; /** - * Blur Threshold - * @description Threshold used in computing noise (lower is better/slower) - * @default 0.2 + * Type + * @default lora + * @constant */ - blur_threshold?: number; + type: "lora"; /** - * Sigma Red - * @description Sigma for strong gaussian blur LPF for red/green - * @default 3 + * Trigger Phrases + * @description Set of trigger phrases for this model */ - sigma_red?: number; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * Sigma Blue - * @description Sigma for weak gaussian blur HPF for blue/green - * @default 1 + * Format + * @default lycoris + * @constant */ - sigma_blue?: number; + format: "lycoris"; /** - * type - * @default noise_spectral + * Base + * @default z-image * @constant */ - type: "noise_spectral"; + base: "z-image"; + variant: components["schemas"]["ZImageVariantType"] | null; }; - /** - * Normal Map - * @description Generates a normal map. - */ - NormalMapInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + /** LoRA_OMI_FLUX_Config */ + LoRA_OMI_FLUX_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * @description The image to process - * @default null + * File Size + * @description The size of the model in bytes. */ - image?: components["schemas"]["ImageField"] | null; + file_size: number; /** - * type - * @default normal_map - * @constant + * Name + * @description Name of the model. */ - type: "normal_map"; - }; - /** - * Octree Quantizer - * @description Quantizes an image to the desired number of colors - */ - OctreeQuantizerInvocation: { + name: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Description + * @description Model description */ - metadata?: components["schemas"]["MetadataField"] | null; + description: string | null; /** - * @description The board to save the image to - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - board?: components["schemas"]["BoardField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - id: string; + source_api_response: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Cover Image + * @description Url for image to preview model */ - is_intermediate?: boolean; + cover_image: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Type + * @default lora + * @constant */ - use_cache?: boolean; + type: "lora"; /** - * @description The image to quantize - * @default null + * Trigger Phrases + * @description Set of trigger phrases for this model */ - image?: components["schemas"]["ImageField"] | null; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * Final Colors - * @description The final number of colors in the palette - * @default 16 + * Format + * @default omi + * @constant */ - final_colors?: number; + format: "omi"; /** - * type - * @default octree_quantizer + * Base + * @default flux * @constant */ - type: "octree_quantizer"; + base: "flux"; }; - /** - * Offset Latents - * @description Offsets a latents tensor by a given percentage of height/width. - * - * This takes a Latents input as well as two numbers (between 0 and 1), which are used to offset the latents in the vertical and/or horizontal directions. 0.5/0.5 would offset the image 50% in both directions such that the corners will wrap around and become the center of the image. - */ - OffsetLatentsInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + /** LoRA_OMI_SDXL_Config */ + LoRA_OMI_SDXL_Config: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Key + * @description A unique key for this model. */ - use_cache?: boolean; + key: string; /** - * @description Latents tensor - * @default null + * Hash + * @description The hash of the model file(s). */ - latents?: components["schemas"]["LatentsField"] | null; + hash: string; /** - * X Offset - * @description Approx percentage to offset (H) - * @default 0.5 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - x_offset?: number; + path: string; /** - * Y Offset - * @description Approx percentage to offset (V) - * @default 0.5 + * File Size + * @description The size of the model in bytes. */ - y_offset?: number; + file_size: number; /** - * type - * @default offset_latents - * @constant + * Name + * @description Name of the model. */ - type: "offset_latents"; - }; - /** OffsetPaginatedResults[BoardDTO] */ - OffsetPaginatedResults_BoardDTO_: { + name: string; /** - * Limit - * @description Limit of items to get + * Description + * @description Model description */ - limit: number; + description: string | null; /** - * Offset - * @description Offset from which to retrieve items + * Source + * @description The original source of the model (path, URL or repo_id). */ - offset: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Total - * @description Total number of items in result + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - total: number; + source_api_response: string | null; /** - * Items - * @description Items + * Cover Image + * @description Url for image to preview model */ - items: components["schemas"]["BoardDTO"][]; - }; - /** OffsetPaginatedResults[ImageDTO] */ - OffsetPaginatedResults_ImageDTO_: { + cover_image: string | null; /** - * Limit - * @description Limit of items to get + * Type + * @default lora + * @constant */ - limit: number; + type: "lora"; /** - * Offset - * @description Offset from which to retrieve items + * Trigger Phrases + * @description Set of trigger phrases for this model */ - offset: number; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["LoraModelDefaultSettings"] | null; /** - * Total - * @description Total number of items in result + * Format + * @default omi + * @constant */ - total: number; + format: "omi"; /** - * Items - * @description Items + * Base + * @default sdxl + * @constant */ - items: components["schemas"]["ImageDTO"][]; + base: "sdxl"; }; /** - * Optimized Tile Size From Area - * @description Cuts up an image into overlapping tiles, optimized for the constraints of min area, max area, overlap, and - * maximum tile dimension (either width or height). Returns ideal width and height. + * LocalModelSource + * @description A local file or directory path. */ - OptimizedTileSizeFromAreaInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + LocalModelSource: { + /** Path */ + path: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Inplace * @default false */ - is_intermediate?: boolean; + inplace?: boolean | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * @description discriminator enum property added by openapi-typescript + * @enum {string} */ - use_cache?: boolean; + type: "local"; + }; + /** + * LogLevel + * @enum {integer} + */ + LogLevel: 0 | 10 | 20 | 30 | 40 | 50; + /** + * LoginRequest + * @description Request body for user login. + */ + LoginRequest: { /** - * Width - * @description Image width in pixels - * @default null + * Email + * @description User email address */ - width?: number | null; + email: string; /** - * Height - * @description Image height in pixels - * @default null + * Password + * @description User password */ - height?: number | null; + password: string; /** - * Min Area - * @description Minimum tile area - * @default 262144 + * Remember Me + * @description Whether to extend session duration + * @default false */ - min_area?: number; + remember_me?: boolean; + }; + /** + * LoginResponse + * @description Response from successful login. + */ + LoginResponse: { /** - * Max Area - * @description Maximum tile area - * @default 1048576 + * Token + * @description JWT access token */ - max_area?: number; + token: string; + /** @description User information */ + user: components["schemas"]["UserDTO"]; /** - * Overlap - * @description Minimum overlap in pixels (both directions) - * @default null + * Expires In + * @description Token expiration time in seconds */ - overlap?: number | null; + expires_in: number; + }; + /** + * LogoutResponse + * @description Response from logout. + */ + LogoutResponse: { /** - * Max Tile Dim - * @description Maximum dimension of a tile (to limit artifacts) - * @default 2048 + * Success + * @description Whether logout was successful */ - max_tile_dim?: number; + success: boolean; + }; + /** LoraModelDefaultSettings */ + LoraModelDefaultSettings: { /** - * type - * @default optimized_tile_size_from_area - * @constant + * Weight + * @description Default weight for this model */ - type: "optimized_tile_size_from_area"; + weight?: number | null; }; - /** - * OrphanedModelInfo - * @description Information about an orphaned model directory. - */ - OrphanedModelInfo: { + /** MDControlListOutput */ + MDControlListOutput: { /** - * Path - * @description Relative path to the orphaned directory from models root + * ControlNet-List + * @description ControlNet(s) to apply */ - path: string; + control_list: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; /** - * Absolute Path - * @description Absolute path to the orphaned directory + * type + * @default md_control_list_output + * @constant */ - absolute_path: string; + type: "md_control_list_output"; + }; + /** MDIPAdapterListOutput */ + MDIPAdapterListOutput: { /** - * Files - * @description List of model files in this directory + * IP-Adapter-List + * @description IP-Adapter to apply */ - files: string[]; + ip_adapter_list: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** - * Size Bytes - * @description Total size of all files in bytes + * type + * @default md_ip_adapter_list_output + * @constant */ - size_bytes: number; + type: "md_ip_adapter_list_output"; }; - /** - * OutputFieldJSONSchemaExtra - * @description Extra attributes to be added to input fields and their OpenAPI schema. Used by the workflow editor - * during schema parsing and UI rendering. - */ - OutputFieldJSONSchemaExtra: { - field_kind: components["schemas"]["FieldKind"]; + /** MDT2IAdapterListOutput */ + MDT2IAdapterListOutput: { /** - * Ui Hidden - * @default false + * T2I Adapter-List + * @description T2I-Adapter(s) to apply */ - ui_hidden: boolean; + t2i_adapter_list: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; /** - * Ui Order - * @default null + * type + * @default md_ip_adapters_output + * @constant */ - ui_order: number | null; - /** @default null */ - ui_type: components["schemas"]["UIType"] | null; + type: "md_ip_adapters_output"; }; /** - * PBR Maps - * @description Generate Normal, Displacement and Roughness Map from a given image + * MLSD Detection + * @description Generates an line segment map using MLSD. */ - PBRMapsInvocation: { + MLSDDetectionInvocation: { /** * @description The board to save the image to * @default null @@ -32397,59 +17415,87 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image + * @description The image to process * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Tile Size - * @description Tile size - * @default 512 + * Score Threshold + * @description The threshold used to score points when determining line segments + * @default 0.1 */ - tile_size?: number; + score_threshold?: number; /** - * Border Mode - * @description Border mode to apply to eliminate any artifacts or seams - * @default none - * @enum {string} + * Distance Threshold + * @description Threshold for including a line segment - lines shorter than this distance will be discarded + * @default 20 */ - border_mode?: "none" | "seamless" | "mirror" | "replicate"; + distance_threshold?: number; + /** + * type + * @default mlsd_detection + * @constant + */ + type: "mlsd_detection"; + }; + /** MainModelDefaultSettings */ + MainModelDefaultSettings: { + /** + * Vae + * @description Default VAE for this model (model key) + */ + vae?: string | null; + /** + * Vae Precision + * @description Default VAE precision for this model + */ + vae_precision?: ("fp16" | "fp32") | null; + /** + * Scheduler + * @description Default scheduler for this model + */ + scheduler?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; + /** + * Steps + * @description Default number of steps for this model + */ + steps?: number | null; + /** + * Cfg Scale + * @description Default CFG Scale for this model + */ + cfg_scale?: number | null; /** - * type - * @default pbr_maps - * @constant + * Cfg Rescale Multiplier + * @description Default CFG Rescale Multiplier for this model */ - type: "pbr_maps"; - }; - /** PBRMapsOutput */ - PBRMapsOutput: { + cfg_rescale_multiplier?: number | null; /** - * @description The generated normal map - * @default null + * Width + * @description Default width for this model */ - normal_map: components["schemas"]["ImageField"]; + width?: number | null; /** - * @description The generated roughness map - * @default null + * Height + * @description Default height for this model */ - roughness_map: components["schemas"]["ImageField"]; + height?: number | null; /** - * @description The generated displacement map - * @default null + * Guidance + * @description Default Guidance for this model */ - displacement_map: components["schemas"]["ImageField"]; + guidance?: number | null; /** - * type - * @default pbr_maps-output - * @constant + * Cpu Only + * @description Whether this model should run on CPU only */ - type: "pbr_maps-output"; + cpu_only?: boolean | null; }; /** - * PTFields Collect - * @description Collect Prompt Tools Fields for an image generated in InvokeAI. + * Main Model - SD1.5, SD2 + * @description Loads a main model, outputting its submodels. */ - PTFieldsCollectInvocation: { + MainModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -32468,1429 +17514,1325 @@ export type components = { */ use_cache?: boolean; /** - * Positive Prompt - * @description The positive prompt parameter + * @description Main model (UNet, VAE, CLIP) to load * @default null */ - positive_prompt?: string | null; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Positive Style Prompt - * @description The positive style prompt parameter - * @default null + * type + * @default main_model_loader + * @constant */ - positive_style_prompt?: string | null; + type: "main_model_loader"; + }; + /** + * Main_BnBNF4_FLUX_Config + * @description Model config for main checkpoint models. + */ + Main_BnBNF4_FLUX_Config: { /** - * Negative Prompt - * @description The negative prompt parameter - * @default null + * Key + * @description A unique key for this model. */ - negative_prompt?: string | null; + key: string; /** - * Negative Style Prompt - * @description The negative prompt parameter - * @default null + * Hash + * @description The hash of the model file(s). */ - negative_style_prompt?: string | null; + hash: string; /** - * Seed - * @description Seed for random number generation - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - seed?: number | null; + path: string; /** - * Width - * @description Width of output (px) - * @default null + * File Size + * @description The size of the model in bytes. */ - width?: number | null; + file_size: number; /** - * Height - * @description Height of output (px) - * @default null + * Name + * @description Name of the model. */ - height?: number | null; + name: string; /** - * Steps - * @description Number of steps to run - * @default null + * Description + * @description Model description */ - steps?: number | null; + description: string | null; /** - * Cfg Scale - * @description Classifier-Free Guidance scale - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - cfg_scale?: number | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - denoising_start?: number | null; + source_api_response: string | null; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default null + * Cover Image + * @description Url for image to preview model */ - denoising_end?: number | null; + cover_image: string | null; /** - * Scheduler - * @description Scheduler to use during inference - * @default null + * Type + * @default main + * @constant */ - scheduler?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; + type: "main"; /** - * type - * @default pt_fields_collect - * @constant + * Trigger Phrases + * @description Set of trigger phrases for this model */ - type: "pt_fields_collect"; - }; - /** - * PTFieldsCollectOutput - * @description PTFieldsCollect Output - */ - PTFieldsCollectOutput: { + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Pt Fields - * @description PTFields in Json Format + * Config Path + * @description Path to the config for this model, if any. */ - pt_fields: string; + config_path: string | null; /** - * type - * @default pt_fields_collect_output + * Base + * @default flux * @constant */ - type: "pt_fields_collect_output"; + base: "flux"; + /** + * Format + * @default bnb_quantized_nf4b + * @constant + */ + format: "bnb_quantized_nf4b"; + variant: components["schemas"]["FluxVariantType"]; }; /** - * PTFields Expand - * @description Save Expand PTFields into individual items + * Main_Checkpoint_FLUX_Config + * @description Model config for main checkpoint models. */ - PTFieldsExpandInvocation: { + Main_Checkpoint_FLUX_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Pt Fields - * @description PTFields in json Format - * @default null + * File Size + * @description The size of the model in bytes. */ - pt_fields?: string | null; + file_size: number; /** - * type - * @default pt_fields_expand + * Name + * @description Name of the model. + */ + name: string; + /** + * Description + * @description Model description + */ + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default main * @constant */ - type: "pt_fields_expand"; - }; - /** - * PTFieldsExpandOutput - * @description Expand Prompt Tools Fields for an image generated in InvokeAI. - */ - PTFieldsExpandOutput: { + type: "main"; /** - * Positive Prompt - * @description The positive prompt + * Trigger Phrases + * @description Set of trigger phrases for this model */ - positive_prompt: string; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Positive Style Prompt - * @description The positive style prompt + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Format + * @default checkpoint + * @constant */ - positive_style_prompt: string; + format: "checkpoint"; /** - * Negative Prompt - * @description The negative prompt + * Base + * @default flux + * @constant */ - negative_prompt: string; + base: "flux"; + variant: components["schemas"]["FluxVariantType"]; + }; + /** + * Main_Checkpoint_Flux2_Config + * @description Model config for FLUX.2 checkpoint models (e.g. Klein). + */ + Main_Checkpoint_Flux2_Config: { /** - * Negative Style Prompt - * @description The negative prompt + * Key + * @description A unique key for this model. */ - negative_style_prompt: string; + key: string; /** - * Seed - * @description Seed for random number generation + * Hash + * @description The hash of the model file(s). */ - seed: number; + hash: string; /** - * Width - * @description Width of output (px) + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - width: number; + path: string; /** - * Height - * @description Height of output (px) + * File Size + * @description The size of the model in bytes. */ - height: number; + file_size: number; /** - * Steps - * @description Number of steps to run + * Name + * @description Name of the model. */ - steps: number; + name: string; /** - * Cfg Scale - * @description Classifier-Free Guidance scale + * Description + * @description Model description */ - cfg_scale: number; + description: string | null; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps + * Source + * @description The original source of the model (path, URL or repo_id). */ - denoising_start: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - denoising_end: number; + source_api_response: string | null; /** - * Scheduler - * @description Scheduler to use during inference - * @enum {string} + * Cover Image + * @description Url for image to preview model */ - scheduler: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + cover_image: string | null; /** - * type - * @default pt_fields_expand_output + * Type + * @default main * @constant */ - type: "pt_fields_expand_output"; - }; - /** PaginatedResults[WorkflowRecordListItemWithThumbnailDTO] */ - PaginatedResults_WorkflowRecordListItemWithThumbnailDTO_: { - /** - * Page - * @description Current Page - */ - page: number; + type: "main"; /** - * Pages - * @description Total number of pages + * Trigger Phrases + * @description Set of trigger phrases for this model */ - pages: number; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Per Page - * @description Number of items per page + * Config Path + * @description Path to the config for this model, if any. */ - per_page: number; + config_path: string | null; /** - * Total - * @description Total number of items in result + * Format + * @default checkpoint + * @constant */ - total: number; + format: "checkpoint"; /** - * Items - * @description Items + * Base + * @default flux2 + * @constant */ - items: components["schemas"]["WorkflowRecordListItemWithThumbnailDTO"][]; + base: "flux2"; + variant: components["schemas"]["Flux2VariantType"]; }; - /** - * Pair Tile with Image - * @description Pair an image with its tile properties. - */ - PairTileImageInvocation: { + /** Main_Checkpoint_SD1_Config */ + Main_Checkpoint_SD1_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * @description The tile image. - * @default null + * File Size + * @description The size of the model in bytes. */ - image?: components["schemas"]["ImageField"] | null; + file_size: number; /** - * @description The tile properties. - * @default null + * Name + * @description Name of the model. */ - tile?: components["schemas"]["Tile"] | null; + name: string; /** - * type - * @default pair_tile_image - * @constant + * Description + * @description Model description */ - type: "pair_tile_image"; - }; - /** PairTileImageOutput */ - PairTileImageOutput: { - /** @description A tile description with its corresponding image. */ - tile_with_image: components["schemas"]["TileWithImage"]; + description: string | null; /** - * type - * @default pair_tile_image_output - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "pair_tile_image_output"; - }; - /** - * PaletteOutput - * @description Base class for Cell Fracture output - */ - PaletteOutput: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * @description The palette output - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - image?: components["schemas"]["ImageField"]; + source_api_response: string | null; /** - * type - * @default get_palette_output - * @constant + * Cover Image + * @description Url for image to preview model */ - type: "get_palette_output"; - }; - /** - * Parse Weighted String - * @description Parses a string containing weighted terms (e.g. `(word)++` or `word-`) and returns the cleaned string, list of terms, their weights, and positions. - */ - ParseWeightedStringInvocation: { + cover_image: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Type + * @default main + * @constant */ - id: string; + type: "main"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Trigger Phrases + * @description Set of trigger phrases for this model */ - is_intermediate?: boolean; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Config Path + * @description Path to the config for this model, if any. */ - use_cache?: boolean; + config_path: string | null; /** - * Text - * @description The input string containing weighted expressions - * @default null + * Format + * @default checkpoint + * @constant */ - text?: string | null; + format: "checkpoint"; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * type - * @default parse_weighted_string + * Base + * @default sd-1 * @constant */ - type: "parse_weighted_string"; + base: "sd-1"; }; - /** - * Paste Image into Bounding Box - * @description Paste the source image into the target image at the given bounding box. - * - * The source image must be the same size as the bounding box, and the bounding box must fit within the target image. - */ - PasteImageIntoBoundingBoxInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + /** Main_Checkpoint_SD2_Config */ + Main_Checkpoint_SD2_Config: { /** - * @description Optional metadata to be saved with the image - * @default null + * Key + * @description A unique key for this model. */ - metadata?: components["schemas"]["MetadataField"] | null; + key: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Hash + * @description The hash of the model file(s). */ - id: string; + hash: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - is_intermediate?: boolean; + path: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * File Size + * @description The size of the model in bytes. */ - use_cache?: boolean; + file_size: number; /** - * @description The image to paste - * @default null + * Name + * @description Name of the model. */ - source_image?: components["schemas"]["ImageField"] | null; + name: string; /** - * @description The image to paste into - * @default null + * Description + * @description Model description */ - target_image?: components["schemas"]["ImageField"] | null; + description: string | null; /** - * @description The bounding box to paste the image into - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - bounding_box?: components["schemas"]["BoundingBoxField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * type - * @default paste_image_into_bounding_box - * @constant - */ - type: "paste_image_into_bounding_box"; - }; - /** - * Percent To Float - * @description Converts a string to a float and divides it by 100. - */ - PercentToFloatInvocation: { + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Cover Image + * @description Url for image to preview model */ - id: string; + cover_image: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Type + * @default main + * @constant */ - is_intermediate?: boolean; + type: "main"; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Trigger Phrases + * @description Set of trigger phrases for this model */ - use_cache?: boolean; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Text - * @description Input text - * @default null + * Config Path + * @description Path to the config for this model, if any. */ - text?: string | null; + config_path: string | null; /** - * type - * @default percent_to_float + * Format + * @default checkpoint * @constant */ - type: "percent_to_float"; - }; - /** - * PiDiNet Edge Detection - * @description Generates an edge map using PiDiNet. - */ - PiDiNetEdgeDetectionInvocation: { + format: "checkpoint"; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * @description The board to save the image to - * @default null + * Base + * @default sd-2 + * @constant */ - board?: components["schemas"]["BoardField"] | null; + base: "sd-2"; + }; + /** Main_Checkpoint_SDXLRefiner_Config */ + Main_Checkpoint_SDXLRefiner_Config: { /** - * @description Optional metadata to be saved with the image - * @default null + * Key + * @description A unique key for this model. */ - metadata?: components["schemas"]["MetadataField"] | null; + key: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Hash + * @description The hash of the model file(s). */ - id: string; + hash: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - is_intermediate?: boolean; + path: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * File Size + * @description The size of the model in bytes. */ - use_cache?: boolean; + file_size: number; /** - * @description The image to process - * @default null + * Name + * @description Name of the model. */ - image?: components["schemas"]["ImageField"] | null; + name: string; /** - * Quantize Edges - * @description Whether or not to use safe mode - * @default false + * Description + * @description Model description */ - quantize_edges?: boolean; + description: string | null; /** - * Scribble - * @description Whether or not to use scribble mode - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - scribble?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * type - * @default pidi_edge_detection - * @constant + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - type: "pidi_edge_detection"; - }; - /** - * Pixel Art - * @description Convert an image to pixelart - */ - PixelArtInvocation: { + source_api_response: string | null; /** - * @description The board to save the image to - * @default null + * Cover Image + * @description Url for image to preview model */ - board?: components["schemas"]["BoardField"] | null; + cover_image: string | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Type + * @default main + * @constant */ - metadata?: components["schemas"]["MetadataField"] | null; + type: "main"; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Trigger Phrases + * @description Set of trigger phrases for this model */ - id: string; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Config Path + * @description Path to the config for this model, if any. */ - is_intermediate?: boolean; + config_path: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Format + * @default checkpoint + * @constant */ - use_cache?: boolean; + format: "checkpoint"; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * @description The image to convert to pixelart - * @default null + * Base + * @default sdxl-refiner + * @constant */ - image?: components["schemas"]["ImageField"] | null; + base: "sdxl-refiner"; + }; + /** Main_Checkpoint_SDXL_Config */ + Main_Checkpoint_SDXL_Config: { /** - * @description Optional - Image with the palette to use - * @default null + * Key + * @description A unique key for this model. */ - palette_image?: components["schemas"]["ImageField"] | null; + key: string; /** - * Pixel Size - * @description Size of the large 'pixels' in the output - * @default 4 + * Hash + * @description The hash of the model file(s). */ - pixel_size?: number; + hash: string; /** - * Palette Size - * @description No. of colours in Palette (used if palette_image is not provided or to extract from it) - * @default 16 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - palette_size?: number; + path: string; /** - * Dither - * @description Apply Floyd-Steinberg dithering - * @default false + * File Size + * @description The size of the model in bytes. */ - dither?: boolean; + file_size: number; /** - * type - * @default pixel-art - * @constant + * Name + * @description Name of the model. */ - type: "pixel-art"; - }; - /** - * Pixelize Image - * @description Creates 'pixel' 'art' using trained models - */ - PixelizeImageInvocation: { + name: string; /** - * @description The board to save the image to - * @default null + * Description + * @description Model description */ - board?: components["schemas"]["BoardField"] | null; + description: string | null; /** - * @description Optional metadata to be saved with the image - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - metadata?: components["schemas"]["MetadataField"] | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - id: string; + source_api_response: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Cover Image + * @description Url for image to preview model */ - is_intermediate?: boolean; + cover_image: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Type + * @default main + * @constant */ - use_cache?: boolean; + type: "main"; /** - * @description The image to pixelize - * @default null + * Trigger Phrases + * @description Set of trigger phrases for this model */ - image?: components["schemas"]["ImageField"] | null; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Correct Colors - * @description Correct hue and saturation to match original image. - * @default true + * Config Path + * @description Path to the config for this model, if any. */ - correct_colors?: boolean; + config_path: string | null; /** - * Cell Size - * @description pixel/cell size (min 2 max WHATEVER BRO) - * @default 4 + * Format + * @default checkpoint + * @constant */ - cell_size?: number; + format: "checkpoint"; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * type - * @default pixelize + * Base + * @default sdxl * @constant */ - type: "pixelize"; + base: "sdxl"; }; /** - * Pixelize - * @description Pixelize an image. Downsample, upsample. + * Main_Checkpoint_ZImage_Config + * @description Model config for Z-Image single-file checkpoint models (safetensors, etc). */ - PixelizeInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + Main_Checkpoint_ZImage_Config: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Key + * @description A unique key for this model. */ - is_intermediate?: boolean; + key: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Hash + * @description The hash of the model file(s). */ - use_cache?: boolean; + hash: string; /** - * @description Input image for pixelization - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - image?: components["schemas"]["ImageField"] | null; + path: string; /** - * Downsample Factor - * @description Image resizing factor. Higher = smaller image. - * @default 4 + * File Size + * @description The size of the model in bytes. */ - downsample_factor?: number; + file_size: number; /** - * Upsample - * @description Upsample to original resolution - * @default true + * Name + * @description Name of the model. */ - upsample?: boolean; + name: string; /** - * type - * @default retro_pixelize - * @constant + * Description + * @description Model description */ - type: "retro_pixelize"; - }; - /** PresetData */ - PresetData: { + description: string | null; /** - * Positive Prompt - * @description Positive prompt + * Source + * @description The original source of the model (path, URL or repo_id). */ - positive_prompt: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Negative Prompt - * @description Negative prompt + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - negative_prompt: string; - }; - /** - * PresetType - * @enum {string} - */ - PresetType: "user" | "default"; - /** - * Print String to Console - * @description Prints a string to the console. - */ - PrintStringToConsoleInvocation: { + source_api_response: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Cover Image + * @description Url for image to preview model */ - id: string; + cover_image: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Type + * @default main + * @constant */ - is_intermediate?: boolean; + type: "main"; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Trigger Phrases + * @description Set of trigger phrases for this model */ - use_cache?: boolean; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Print Colour - * @description The colour to print the console output - * @default White on Black - * @enum {string} + * Config Path + * @description Path to the config for this model, if any. */ - print_colour?: "Black on White" | "White on Black" | "Black on Green" | "Red on White" | "Green on Red" | "Yellow on Blue" | "Blue on Yellow" | "Magenta on Cyan" | "Cyan on Magenta" | "Black on Red" | "Red on Black" | "Yellow on Black" | "Black on Yellow"; + config_path: string | null; /** - * Input string - * @description The string to print to console. - * @default , + * Base + * @default z-image + * @constant */ - input_str?: string; + base: "z-image"; /** - * type - * @default print_string_to_console_invocation + * Format + * @default checkpoint * @constant */ - type: "print_string_to_console_invocation"; + format: "checkpoint"; + variant: components["schemas"]["ZImageVariantType"]; }; - /** - * PrintStringToConsoleOutput - * @description Debug Print String Output - */ - PrintStringToConsoleOutput: { + /** Main_Diffusers_CogView4_Config */ + Main_Diffusers_CogView4_Config: { /** - * Passthrough - * @description Passthrough + * Key + * @description A unique key for this model. */ - passthrough: string; + key: string; /** - * type - * @default print_string_to_console_output - * @constant + * Hash + * @description The hash of the model file(s). */ - type: "print_string_to_console_output"; - }; - /** - * ProgressImage - * @description The progress image sent intermittently during processing - */ - ProgressImage: { + hash: string; /** - * Width - * @description The effective width of the image in pixels + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - width: number; + path: string; /** - * Height - * @description The effective height of the image in pixels + * File Size + * @description The size of the model in bytes. */ - height: number; + file_size: number; /** - * Dataurl - * @description The image data as a b64 data URL + * Name + * @description Name of the model. */ - dataURL: string; - }; - /** - * Prompt Auto .and() - * @description Takes a prompt string then chunks it up into a .and() output if over the max length - */ - PromptAutoAndInvocation: { + name: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Description + * @description Model description */ - id: string; + description: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - is_intermediate?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - use_cache?: boolean; + source_api_response: string | null; /** - * Prompt - * @description Prompt to auto .and() - * @default + * Cover Image + * @description Url for image to preview model */ - prompt?: string; + cover_image: string | null; /** - * Max Length - * @description Maximum chunk length in characters - * @default 200 + * Type + * @default main + * @constant */ - max_length?: number; + type: "main"; /** - * type - * @default prompt_auto_and + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + /** + * Base + * @default cogview4 * @constant */ - type: "prompt_auto_and"; + base: "cogview4"; }; /** - * Prompt from Lookup Table - * @description Creates prompts using lookup table templates + * Main_Diffusers_FLUX_Config + * @description Model config for FLUX.1 models in diffusers format. */ - PromptFromLookupTableInvocation: { + Main_Diffusers_FLUX_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * Resolutions Dict - * @description Private field for id substitutions dict cache - * @default {} + * File Size + * @description The size of the model in bytes. */ - resolutions_dict?: { - [key: string]: unknown; - }; + file_size: number; /** - * Lookups - * @description Lookup table(s) containing template(s) (JSON) - * @default [] + * Name + * @description Name of the model. */ - lookups?: string | string[]; + name: string; /** - * Remove Negatives - * @description Whether to strip out text between [] - * @default false + * Description + * @description Model description */ - remove_negatives?: boolean; + description: string | null; /** - * Strip Parens Probability - * @description Probability of removing attention group weightings - * @default 0 + * Source + * @description The original source of the model (path, URL or repo_id). */ - strip_parens_probability?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Resolutions - * @description JSON structure of substitutions by id by tag - * @default [] + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - resolutions?: string | string[]; + source_api_response: string | null; /** - * Seed Vector In - * @description Optional JSON array of seeds for deterministic generation - * @default null + * Cover Image + * @description Url for image to preview model */ - seed_vector_in?: string | null; + cover_image: string | null; /** - * type - * @default prompt_from_lookup_table + * Type + * @default main * @constant */ - type: "prompt_from_lookup_table"; - }; - /** - * Prompt Strength - * @description Takes a prompt string and float strength and outputs a new string in the format of (prompt)strength - */ - PromptStrengthInvocation: { + type: "main"; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Trigger Phrases + * @description Set of trigger phrases for this model */ - id: string; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Format + * @default diffusers + * @constant */ - is_intermediate?: boolean; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Base + * @default flux + * @constant */ - use_cache?: boolean; + base: "flux"; + variant: components["schemas"]["FluxVariantType"]; + }; + /** + * Main_Diffusers_Flux2_Config + * @description Model config for FLUX.2 models in diffusers format (e.g. FLUX.2 Klein). + */ + Main_Diffusers_Flux2_Config: { /** - * Collection - * @description Collection of Prompt strengths - * @default [] + * Key + * @description A unique key for this model. */ - collection?: string[]; + key: string; /** - * Prompt - * @description Prompt to work on - * @default + * Hash + * @description The hash of the model file(s). */ - prompt?: string; + hash: string; /** - * Strength - * @description strength of the prompt - * @default 1 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - strength?: number; + path: string; /** - * type - * @default prompt_strength - * @constant + * File Size + * @description The size of the model in bytes. */ - type: "prompt_strength"; - }; - /** - * PromptStrengthOutput - * @description Base class for nodes that output a collection of images - */ - PromptStrengthOutput: { + file_size: number; /** - * Collection - * @description Prompt strength collection + * Name + * @description Name of the model. */ - collection: string[]; + name: string; /** - * Value - * @description Prompt strength + * Description + * @description Model description */ - value: string; + description: string | null; /** - * type - * @default prompt_strength_collection_output - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "prompt_strength_collection_output"; - }; - /** - * Prompt Strengths Combine - * @description Takes a collection of prompt strength strings and converts it into a combined .and() or .blend() structure. Blank prompts are ignored - */ - PromptStrengthsCombineInvocation: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - id: string; + source_api_response: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Cover Image + * @description Url for image to preview model */ - is_intermediate?: boolean; + cover_image: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Type + * @default main + * @constant */ - use_cache?: boolean; + type: "main"; /** - * Prompt Strengths - * @description Prompt strengths to combine - * @default [ - * "" - * ] + * Trigger Phrases + * @description Set of trigger phrases for this model */ - prompt_strengths?: string[]; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Combine Type - * @description Combine type .and() or .blend() - * @default .and - * @enum {string} + * Format + * @default diffusers + * @constant */ - combine_type?: ".and" | ".blend"; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * type - * @default prompt_strengths_combine + * Base + * @default flux2 * @constant */ - type: "prompt_strengths_combine"; + base: "flux2"; + variant: components["schemas"]["Flux2VariantType"]; }; - /** - * Prompt Template - * @description Applies a Style Preset template to positive and negative prompts. - * - * Select a Style Preset and provide positive/negative prompts. The node replaces - * {prompt} placeholders in the template with your input prompts. - */ - PromptTemplateInvocation: { + /** Main_Diffusers_SD1_Config */ + Main_Diffusers_SD1_Config: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description A unique key for this model. */ - id: string; + key: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description The hash of the model file(s). */ - is_intermediate?: boolean; + hash: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - use_cache?: boolean; + path: string; /** - * @description The Style Preset to use as a template - * @default null + * File Size + * @description The size of the model in bytes. */ - style_preset?: components["schemas"]["StylePresetField"] | null; + file_size: number; /** - * Positive Prompt - * @description The positive prompt to insert into the template's {prompt} placeholder - * @default + * Name + * @description Name of the model. */ - positive_prompt?: string; + name: string; /** - * Negative Prompt - * @description The negative prompt to insert into the template's {prompt} placeholder - * @default + * Description + * @description Model description */ - negative_prompt?: string; + description: string | null; /** - * type - * @default prompt_template - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "prompt_template"; - }; - /** - * PromptTemplateOutput - * @description Output for the Prompt Template node - */ - PromptTemplateOutput: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Positive Prompt - * @description The positive prompt with the template applied + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - positive_prompt: string; + source_api_response: string | null; /** - * Negative Prompt - * @description The negative prompt with the template applied + * Cover Image + * @description Url for image to preview model */ - negative_prompt: string; + cover_image: string | null; /** - * type - * @default prompt_template_output + * Type + * @default main * @constant */ - type: "prompt_template_output"; - }; - /** - * Prompts from File - * @description Loads prompts from a text file - */ - PromptsFromFileInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + type: "main"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Trigger Phrases + * @description Set of trigger phrases for this model */ - is_intermediate?: boolean; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Format + * @default diffusers + * @constant */ - use_cache?: boolean; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * File Path - * @description Path to prompt text file - * @default null + * Base + * @default sd-1 + * @constant */ - file_path?: string | null; + base: "sd-1"; + }; + /** Main_Diffusers_SD2_Config */ + Main_Diffusers_SD2_Config: { /** - * Pre Prompt - * @description String to prepend to each prompt - * @default null + * Key + * @description A unique key for this model. */ - pre_prompt?: string | null; + key: string; /** - * Post Prompt - * @description String to append to each prompt - * @default null + * Hash + * @description The hash of the model file(s). */ - post_prompt?: string | null; + hash: string; /** - * Start Line - * @description Line in the file to start start from - * @default 1 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - start_line?: number; + path: string; /** - * Max Prompts - * @description Max lines to read from file (0=all) - * @default 1 + * File Size + * @description The size of the model in bytes. */ - max_prompts?: number; + file_size: number; /** - * type - * @default prompt_from_file - * @constant + * Name + * @description Name of the model. */ - type: "prompt_from_file"; - }; - /** - * Prompts To File - * @description Save prompts to a text file - */ - PromptsToFileInvocation: { + name: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Description + * @description Model description */ - id: string; + description: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - is_intermediate?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - use_cache?: boolean; + source_api_response: string | null; /** - * File Path - * @description Path to prompt text file - * @default null + * Cover Image + * @description Url for image to preview model */ - file_path?: string | null; + cover_image: string | null; /** - * Prompts - * @description Prompt or collection of prompts to write - * @default null + * Type + * @default main + * @constant */ - prompts?: string | string[] | null; + type: "main"; /** - * Append - * @description Append or overwrite file - * @default true + * Trigger Phrases + * @description Set of trigger phrases for this model */ - append?: boolean; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * type - * @default prompt_to_file + * Format + * @default diffusers * @constant */ - type: "prompt_to_file"; - }; - /** - * PromptsToFileInvocationOutput - * @description Base class for invocation that writes to a file and returns nothing of use - */ - PromptsToFileInvocationOutput: { + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * type - * @default prompt_to_file_output + * Base + * @default sd-2 * @constant */ - type: "prompt_to_file_output"; + base: "sd-2"; }; - /** - * PruneResult - * @description Result of pruning the session queue - */ - PruneResult: { + /** Main_Diffusers_SD3_Config */ + Main_Diffusers_SD3_Config: { /** - * Deleted - * @description Number of queue items deleted + * Key + * @description A unique key for this model. */ - deleted: number; - }; - /** - * Prune/Clean Prompts - * @description Like home staging but for prompt text - */ - PruneTextInvocation: { + key: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Hash + * @description The hash of the model file(s). */ - id: string; + hash: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - is_intermediate?: boolean; + path: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * File Size + * @description The size of the model in bytes. */ - use_cache?: boolean; + file_size: number; /** - * Content - * @description Text to prune/cleanup - * @default null + * Name + * @description Name of the model. */ - content?: string | null; + name: string; /** - * Blacklist File Path - * @description Path to .txt to prune with. No path will run without matched content removal. - * @default + * Description + * @description Model description */ - blacklist_file_path?: string | null; + description: string | null; /** - * Remove Weight Syntax - * @description Remove basic Compel + A111-style attention weighting syntax. Special Compel syntax like .and() not supported - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - remove_weight_syntax?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Dedupe Tags - * @description Group text by commas, remove duplicates - * @default true + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - dedupe_tags?: boolean; + source_api_response: string | null; /** - * Remove Slashes - * @description Delete all backslashes instead of just extras - * @default true + * Cover Image + * @description Url for image to preview model */ - remove_slashes?: boolean; + cover_image: string | null; /** - * Remove Tis And Loras - * @description Delete Invoke TI syntax, A111 LoRAs, and Invoke 2.x LoRAs - * @default false + * Type + * @default main + * @constant */ - remove_tis_and_loras?: boolean; + type: "main"; /** - * Custom Regex Pattern - * @description Custom regex pattern to apply to the text - * @default + * Trigger Phrases + * @description Set of trigger phrases for this model */ - custom_regex_pattern?: string | null; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Custom Regex Substitution - * @description Substitution string for the custom regex pattern. Leave blank to simply delete captured text - * @default + * Format + * @default diffusers + * @constant */ - custom_regex_substitution?: string | null; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * type - * @default prune_prompt + * Base + * @default sd-3 * @constant */ - type: "prune_prompt"; + base: "sd-3"; + /** + * Submodels + * @description Loadable submodels in this model + */ + submodels: { + [key: string]: components["schemas"]["SubmodelDefinition"]; + } | null; }; - /** - * PrunedPromptOutput - * @description Pruned and cleaned string output - */ - PrunedPromptOutput: { + /** Main_Diffusers_SDXLRefiner_Config */ + Main_Diffusers_SDXLRefiner_Config: { /** - * Prompt - * @description Processed prompt string + * Key + * @description A unique key for this model. */ - prompt: string; + key: string; /** - * type - * @default clean_prompt - * @constant + * Hash + * @description The hash of the model file(s). */ - type: "clean_prompt"; - }; - /** - * QueueClearedEvent - * @description Event model for queue_cleared - */ - QueueClearedEvent: { + hash: string; /** - * Timestamp - * @description The timestamp of the event + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - timestamp: number; + path: string; /** - * Queue Id - * @description The ID of the queue + * File Size + * @description The size of the model in bytes. */ - queue_id: string; - }; - /** - * QueueItemStatusChangedEvent - * @description Event model for queue_item_status_changed - */ - QueueItemStatusChangedEvent: { + file_size: number; /** - * Timestamp - * @description The timestamp of the event + * Name + * @description Name of the model. */ - timestamp: number; + name: string; /** - * Queue Id - * @description The ID of the queue + * Description + * @description Model description */ - queue_id: string; + description: string | null; /** - * Item Id - * @description The ID of the queue item + * Source + * @description The original source of the model (path, URL or repo_id). */ - item_id: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Batch Id - * @description The ID of the queue batch + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - batch_id: string; + source_api_response: string | null; /** - * Origin - * @description The origin of the queue item - * @default null + * Cover Image + * @description Url for image to preview model */ - origin: string | null; + cover_image: string | null; /** - * Destination - * @description The destination of the queue item - * @default null + * Type + * @default main + * @constant */ - destination: string | null; + type: "main"; /** - * User Id - * @description The ID of the user who created the queue item - * @default system + * Trigger Phrases + * @description Set of trigger phrases for this model */ - user_id: string; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; + /** + * Format + * @default diffusers + * @constant + */ + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; + /** + * Base + * @default sdxl-refiner + * @constant + */ + base: "sdxl-refiner"; + }; + /** Main_Diffusers_SDXL_Config */ + Main_Diffusers_SDXL_Config: { /** - * Status - * @description The new status of the queue item - * @enum {string} + * Key + * @description A unique key for this model. */ - status: "pending" | "in_progress" | "completed" | "failed" | "canceled"; + key: string; /** - * Error Type - * @description The error type, if any - * @default null + * Hash + * @description The hash of the model file(s). */ - error_type: string | null; + hash: string; /** - * Error Message - * @description The error message, if any - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - error_message: string | null; + path: string; /** - * Error Traceback - * @description The error traceback, if any - * @default null + * File Size + * @description The size of the model in bytes. */ - error_traceback: string | null; + file_size: number; /** - * Created At - * @description The timestamp when the queue item was created + * Name + * @description Name of the model. */ - created_at: string; + name: string; /** - * Updated At - * @description The timestamp when the queue item was last updated + * Description + * @description Model description */ - updated_at: string; + description: string | null; /** - * Started At - * @description The timestamp when the queue item was started - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - started_at: string | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Completed At - * @description The timestamp when the queue item was completed - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - completed_at: string | null; - /** @description The status of the batch */ - batch_status: components["schemas"]["BatchStatus"]; - /** @description The status of the queue */ - queue_status: components["schemas"]["SessionQueueStatus"]; + source_api_response: string | null; /** - * Session Id - * @description The ID of the session (aka graph execution state) + * Cover Image + * @description Url for image to preview model */ - session_id: string; - }; - /** - * QueueItemsRetriedEvent - * @description Event model for queue_items_retried - */ - QueueItemsRetriedEvent: { + cover_image: string | null; /** - * Timestamp - * @description The timestamp of the event + * Type + * @default main + * @constant */ - timestamp: number; + type: "main"; /** - * Queue Id - * @description The ID of the queue + * Trigger Phrases + * @description Set of trigger phrases for this model */ - queue_id: string; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Retried Item Ids - * @description The IDs of the queue items that were retried + * Format + * @default diffusers + * @constant */ - retried_item_ids: number[]; - }; - /** - * Qwen3EncoderField - * @description Field for Qwen3 text encoder used by Z-Image models. - */ - Qwen3EncoderField: { - /** @description Info to load tokenizer submodel */ - tokenizer: components["schemas"]["ModelIdentifierField"]; - /** @description Info to load text_encoder submodel */ - text_encoder: components["schemas"]["ModelIdentifierField"]; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; + prediction_type: components["schemas"]["SchedulerPredictionType"]; + variant: components["schemas"]["ModelVariantType"]; /** - * Loras - * @description LoRAs to apply on model loading + * Base + * @default sdxl + * @constant */ - loras?: components["schemas"]["LoRAField"][]; + base: "sdxl"; }; /** - * Qwen3Encoder_Checkpoint_Config - * @description Configuration for single-file Qwen3 Encoder models (safetensors). + * Main_Diffusers_ZImage_Config + * @description Model config for Z-Image diffusers models (Z-Image-Turbo, Z-Image-Base). */ - Qwen3Encoder_Checkpoint_Config: { + Main_Diffusers_ZImage_Config: { /** * Key * @description A unique key for this model. @@ -33939,41 +18881,39 @@ export type components = { */ cover_image: string | null; /** - * Config Path - * @description Path to the config for this model, if any. - */ - config_path: string | null; - /** - * Base - * @default any + * Type + * @default main * @constant */ - base: "any"; + type: "main"; /** - * Type - * @default qwen3_encoder - * @constant + * Trigger Phrases + * @description Set of trigger phrases for this model */ - type: "qwen3_encoder"; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** * Format - * @default checkpoint + * @default diffusers * @constant */ - format: "checkpoint"; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Cpu Only - * @description Whether this model should run on CPU only + * Base + * @default z-image + * @constant */ - cpu_only: boolean | null; - /** @description Qwen3 model size variant (4B or 8B) */ - variant: components["schemas"]["Qwen3VariantType"]; + base: "z-image"; + variant: components["schemas"]["ZImageVariantType"]; }; /** - * Qwen3Encoder_GGUF_Config - * @description Configuration for GGUF-quantized Qwen3 Encoder models. + * Main_GGUF_FLUX_Config + * @description Model config for main checkpoint models. */ - Qwen3Encoder_GGUF_Config: { + Main_GGUF_FLUX_Config: { /** * Key * @description A unique key for this model. @@ -34021,6 +18961,19 @@ export type components = { * @description Url for image to preview model */ cover_image: string | null; + /** + * Type + * @default main + * @constant + */ + type: "main"; + /** + * Trigger Phrases + * @description Set of trigger phrases for this model + */ + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** * Config Path * @description Path to the config for this model, if any. @@ -34028,38 +18981,23 @@ export type components = { config_path: string | null; /** * Base - * @default any - * @constant - */ - base: "any"; - /** - * Type - * @default qwen3_encoder + * @default flux * @constant */ - type: "qwen3_encoder"; + base: "flux"; /** * Format * @default gguf_quantized * @constant */ format: "gguf_quantized"; - /** - * Cpu Only - * @description Whether this model should run on CPU only - */ - cpu_only: boolean | null; - /** @description Qwen3 model size variant (4B or 8B) */ - variant: components["schemas"]["Qwen3VariantType"]; + variant: components["schemas"]["FluxVariantType"]; }; /** - * Qwen3Encoder_Qwen3Encoder_Config - * @description Configuration for Qwen3 Encoder models in a diffusers-like format. - * - * The model weights are expected to be in a folder called text_encoder inside the model directory, - * compatible with Qwen2VLForConditionalGeneration or similar architectures used by Z-Image. + * Main_GGUF_Flux2_Config + * @description Model config for GGUF-quantized FLUX.2 checkpoint models (e.g. Klein). */ - Qwen3Encoder_Qwen3Encoder_Config: { + Main_GGUF_Flux2_Config: { /** * Key * @description A unique key for this model. @@ -34107,159 +19045,127 @@ export type components = { * @description Url for image to preview model */ cover_image: string | null; - /** - * Base - * @default any - * @constant - */ - base: "any"; /** * Type - * @default qwen3_encoder - * @constant - */ - type: "qwen3_encoder"; - /** - * Format - * @default qwen3_encoder + * @default main * @constant */ - format: "qwen3_encoder"; + type: "main"; /** - * Cpu Only - * @description Whether this model should run on CPU only + * Trigger Phrases + * @description Set of trigger phrases for this model */ - cpu_only: boolean | null; - /** @description Qwen3 model size variant (4B or 8B) */ - variant: components["schemas"]["Qwen3VariantType"]; - }; - /** - * Qwen3 Prompt Pro - * @description Use a Qwen3 encoder as a causal LLM to enhance a prompt, then encode it for Z-Image and Flux Klein. - * - * Accepts the same Qwen3 encoder model used by Z-Image and Flux2 Klein text encoder nodes. - * Outputs the enhanced prompt as text plus ready-to-use conditioning for both pipelines. - */ - Qwen3PromptProInvocation: { + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Config Path + * @description Path to the config for this model, if any. */ - id: string; + config_path: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Base + * @default flux2 + * @constant */ - is_intermediate?: boolean; + base: "flux2"; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Format + * @default gguf_quantized + * @constant */ - use_cache?: boolean; + format: "gguf_quantized"; + variant: components["schemas"]["Flux2VariantType"]; + }; + /** + * Main_GGUF_ZImage_Config + * @description Model config for GGUF-quantized Z-Image transformer models. + */ + Main_GGUF_ZImage_Config: { /** - * Prompt - * @description Input text prompt. - * @default + * Key + * @description A unique key for this model. */ - prompt?: string; + key: string; /** - * System Prompt - * @description System prompt that guides prompt enhancement (generation step). - * @default You are an expert prompt writer for AI image generation. Given a brief description, expand it into a detailed, vivid prompt suitable for generating high-quality images. Only output the expanded prompt, nothing else. + * Hash + * @description The hash of the model file(s). */ - system_prompt?: string; + hash: string; /** - * Conditioning System Prompt - * @description System prompt prepended when encoding the enhanced prompt into conditioning. Leave empty for default behavior. - * @default + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - conditioning_system_prompt?: string; + path: string; /** - * Qwen3 Encoder - * @description Qwen3 tokenizer and text encoder - * @default null + * File Size + * @description The size of the model in bytes. */ - qwen3_encoder?: components["schemas"]["Qwen3EncoderField"] | null; + file_size: number; /** - * @description A mask defining the region that the conditioning applies to. - * @default null + * Name + * @description Name of the model. */ - mask?: components["schemas"]["TensorField"] | null; + name: string; /** - * Max Tokens - * @description Maximum number of tokens to generate. - * @default 300 + * Description + * @description Model description */ - max_tokens?: number; + description: string | null; /** - * Temperature - * @description Sampling temperature. Higher = more creative, lower = more focused. - * @default 0.7 + * Source + * @description The original source of the model (path, URL or repo_id). */ - temperature?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Top P - * @description Nucleus sampling threshold. - * @default 0.9 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - top_p?: number; + source_api_response: string | null; /** - * Enable Thinking - * @description Enable Qwen3 thinking mode during generation. Slower but potentially higher quality. - * @default false + * Cover Image + * @description Url for image to preview model */ - enable_thinking?: boolean; + cover_image: string | null; /** - * type - * @default qwen3_prompt_pro + * Type + * @default main * @constant */ - type: "qwen3_prompt_pro"; - }; - /** - * Qwen3PromptProOutput - * @description Output containing the enhanced prompt text and conditioning for Z-Image and Flux Klein. - */ - Qwen3PromptProOutput: { + type: "main"; /** - * Enhanced Prompt - * @description The enhanced prompt text + * Trigger Phrases + * @description Set of trigger phrases for this model */ - enhanced_prompt: string; - /** @description Z-Image conditioning embeddings */ - z_image_conditioning: components["schemas"]["ZImageConditioningField"]; + trigger_phrases: string[] | null; + /** @description Default settings for this model */ + default_settings: components["schemas"]["MainModelDefaultSettings"] | null; /** - * Z Image Prompt - * @description The prompt text encoded for Z-Image conditioning + * Config Path + * @description Path to the config for this model, if any. */ - z_image_prompt: string; - /** @description Flux Klein conditioning embeddings */ - flux_conditioning: components["schemas"]["FluxConditioningField"]; + config_path: string | null; /** - * Flux Prompt - * @description The prompt text encoded for Flux Klein conditioning + * Base + * @default z-image + * @constant */ - flux_prompt: string; + base: "z-image"; /** - * type - * @default qwen3_prompt_pro_output + * Format + * @default gguf_quantized * @constant */ - type: "qwen3_prompt_pro_output"; + format: "gguf_quantized"; + variant: components["schemas"]["ZImageVariantType"]; }; /** - * Qwen3VariantType - * @description Qwen3 text encoder variants based on model size. - * @enum {string} - */ - Qwen3VariantType: "qwen3_4b" | "qwen3_8b"; - /** - * RGB Merge - * @description Merge RGB color channels and alpha + * Combine Masks + * @description Combine two masks together by multiplying them using `PIL.ImageChops.multiply()`. */ - RGBMergeInvocation: { + MaskCombineInvocation: { /** * @description The board to save the image to * @default null @@ -34288,37 +19194,27 @@ export type components = { */ use_cache?: boolean; /** - * @description The red channel - * @default null - */ - r_channel?: components["schemas"]["ImageField"] | null; - /** - * @description The green channel - * @default null - */ - g_channel?: components["schemas"]["ImageField"] | null; - /** - * @description The blue channel + * @description The first mask to combine * @default null */ - b_channel?: components["schemas"]["ImageField"] | null; + mask1?: components["schemas"]["ImageField"] | null; /** - * @description The alpha channel + * @description The second image to combine * @default null */ - alpha_channel?: components["schemas"]["ImageField"] | null; + mask2?: components["schemas"]["ImageField"] | null; /** * type - * @default rgb_merge + * @default mask_combine * @constant */ - type: "rgb_merge"; + type: "mask_combine"; }; /** - * RGB Split - * @description Split an image into RGB color channels and alpha + * Mask Edge + * @description Applies an edge mask to an image */ - RGBSplitInvocation: { + MaskEdgeInvocation: { /** * @description The board to save the image to * @default null @@ -34347,52 +19243,56 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to split into channels + * @description The image to apply the mask to * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * type - * @default rgb_split - * @constant + * Edge Size + * @description The size of the edge + * @default null */ - type: "rgb_split"; - }; - /** - * RGBSplitOutput - * @description Base class for invocations that output three L-mode images (R, G, B) - */ - RGBSplitOutput: { - /** @description Grayscale image of the red channel */ - r_channel: components["schemas"]["ImageField"]; - /** @description Grayscale image of the green channel */ - g_channel: components["schemas"]["ImageField"]; - /** @description Grayscale image of the blue channel */ - b_channel: components["schemas"]["ImageField"]; - /** @description Grayscale image of the alpha channel */ - alpha_channel: components["schemas"]["ImageField"]; + edge_size?: number | null; /** - * Width - * @description The width of the image in pixels + * Edge Blur + * @description The amount of blur on the edge + * @default null */ - width: number; + edge_blur?: number | null; /** - * Height - * @description The height of the image in pixels + * Low Threshold + * @description First threshold for the hysteresis procedure in Canny edge detection + * @default null */ - height: number; + low_threshold?: number | null; + /** + * High Threshold + * @description Second threshold for the hysteresis procedure in Canny edge detection + * @default null + */ + high_threshold?: number | null; /** * type - * @default rgb_split_output + * @default mask_edge * @constant */ - type: "rgb_split_output"; + type: "mask_edge"; }; /** - * Random Float - * @description Outputs a single random float + * Mask from Alpha + * @description Extracts the alpha channel of an image as a mask. */ - RandomFloatInvocation: { + MaskFromAlphaInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -34407,39 +19307,42 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Low - * @description The inclusive low value - * @default 0 - */ - low?: number; - /** - * High - * @description The exclusive high value - * @default 1 + * @description The image to create the mask from + * @default null */ - high?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Decimals - * @description The number of decimal places to round to - * @default 2 + * Invert + * @description Whether or not to invert the mask + * @default false */ - decimals?: number; + invert?: boolean; /** * type - * @default rand_float + * @default tomask * @constant */ - type: "rand_float"; + type: "tomask"; }; /** - * Random Image Size - * @description Outputs a random size from a list of sizes. + * Mask from Segmented Image + * @description Generate a mask for a particular color in an ID Map */ - RandomImageSizeInvocation: { + MaskFromIDInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -34458,90 +19361,73 @@ export type components = { */ use_cache?: boolean; /** - * Seed - * @description A seed for the randomness. Use -1 for non-deterministic. - * @default -1 + * @description The image to create the mask from + * @default null */ - seed?: number; + image?: components["schemas"]["ImageField"] | null; + /** + * @description ID color to mask + * @default null + */ + color?: components["schemas"]["ColorField"] | null; /** - * Size List - * @default 1024x1024,888x1184,1184x888,840x1256,1256x840,768x1368,1368x768 + * Threshold + * @description Threshold for color detection + * @default 100 */ - size_list?: string; + threshold?: number; /** - * Multiples of 32 + * Invert + * @description Whether or not to invert the mask * @default false */ - multiples_of_32?: boolean; + invert?: boolean; /** * type - * @default random_image_size_invocation + * @default mask_from_id * @constant */ - type: "random_image_size_invocation"; + type: "mask_from_id"; }; /** - * RandomImageSizeOutput - * @description Random Image Size Output + * MaskOutput + * @description A torch mask tensor. */ - RandomImageSizeOutput: { - /** Width */ + MaskOutput: { + /** @description The mask. */ + mask: components["schemas"]["TensorField"]; + /** + * Width + * @description The width of the mask in pixels. + */ width: number; - /** Height */ + /** + * Height + * @description The height of the mask in pixels. + */ height: number; /** * type - * @default random_image_size + * @default mask_output * @constant */ - type: "random_image_size"; + type: "mask_output"; }; /** - * Random Integer - * @description Outputs a single random integer. + * Tensor Mask to Image + * @description Convert a mask tensor to an image. */ - RandomIntInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default false - */ - use_cache?: boolean; - /** - * Low - * @description The inclusive low value - * @default 0 - */ - low?: number; + MaskTensorToImageInvocation: { /** - * High - * @description The exclusive high value - * @default 2147483647 + * @description The board to save the image to + * @default null */ - high?: number; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default rand_int - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "rand_int"; - }; - /** - * Random LoRA Mixer - * @description Returns a random Collection of LoRAs selected from an input Collection. - */ - RandomLoRAMixerInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -34556,96 +19442,36 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * LoRAs - * @description A random selection of the input LoRAs with random weights + * @description The mask tensor to convert. * @default null */ - loras?: components["schemas"]["LoRAField"][] | null; - /** - * Seed - * @description A seed to use for the random number generator. This allows reproducability. Leave as -1 for non-deterministic - * @default -1 - */ - seed?: number; - /** - * Min LoRAs - * @description The mininum number of LoRAs to output - * @default 1 - */ - min_loras?: number; - /** - * Max LoRAs - * @description The maximum number of LoRAs to output - * @default 3 - */ - max_loras?: number; - /** - * Min Weight - * @description The mininum weight to apply to a LoRA - * @default 0.05 - */ - min_weight?: number; - /** - * Max Weight - * @description The maxinum weight to apply to a LoRA - * @default 1 - */ - max_weight?: number; - /** - * Trigger Word Delimiter - * @description A character sequence to place between each trigger word - * @default , - */ - trigger_word_delimiter?: string; - /** - * LoRA Names Delimiter - * @description A character sequence to place between each LoRA name - * @default , - */ - lora_names_delimiter?: string; + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default random_lora_mixer_invocation + * @default tensor_mask_to_image * @constant */ - type: "random_lora_mixer_invocation"; + type: "tensor_mask_to_image"; }; /** - * RandomLoRAMixerOutput - * @description Random LoRA Mixer Output + * MediaPipe Face Detection + * @description Detects faces using MediaPipe. */ - RandomLoRAMixerOutput: { - /** - * LoRAs - * @description A random selection of the input LoRAs with random weights - */ - loras: components["schemas"]["LoRAField"][]; - /** - * Trigger Words - * @description A string with delimited trigger words for the random LoRAs - */ - trigger_words: string; + MediaPipeFaceDetectionInvocation: { /** - * Lora Names - * @description A string with the names and weights of all applied LoRAs + * @description The board to save the image to + * @default null */ - lora_names: string; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default random_lora_mixer_output - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "random_lora_mixer_output"; - }; - /** - * Random Range - * @description Creates a collection of random numbers - */ - RandomRangeInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -34660,45 +19486,38 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * Low - * @description The inclusive low value - * @default 0 - */ - low?: number; - /** - * High - * @description The exclusive high value - * @default 2147483647 + * @description The image to process + * @default null */ - high?: number; + image?: components["schemas"]["ImageField"] | null; /** - * Size - * @description The number of values to generate + * Max Faces + * @description Maximum number of faces to detect * @default 1 */ - size?: number; + max_faces?: number; /** - * Seed - * @description The seed for the RNG (omit for random) - * @default 0 + * Min Confidence + * @description Minimum confidence for face detection + * @default 0.5 */ - seed?: number; + min_confidence?: number; /** * type - * @default random_range + * @default mediapipe_face_detection * @constant */ - type: "random_range"; + type: "mediapipe_face_detection"; }; /** - * Integer Range - * @description Creates a range of numbers from start to stop with step + * Metadata Merge + * @description Merged a collection of MetadataDict into a single MetadataDict. */ - RangeInvocation: { + MergeMetadataInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -34717,35 +19536,33 @@ export type components = { */ use_cache?: boolean; /** - * Start - * @description The start of the range - * @default 0 - */ - start?: number; - /** - * Stop - * @description The stop of the range - * @default 10 - */ - stop?: number; - /** - * Step - * @description The step of the range - * @default 1 + * Collection + * @description Collection of Metadata + * @default null */ - step?: number; + collection?: components["schemas"]["MetadataField"][] | null; /** * type - * @default range + * @default merge_metadata * @constant */ - type: "range"; + type: "merge_metadata"; }; /** - * Integer Range of Size - * @description Creates a range from start to start + (size * step) incremented by step + * Merge Tiles to Image + * @description Merge multiple tile images into a single image. */ - RangeOfSizeInvocation: { + MergeTilesToImageInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -34764,35 +19581,43 @@ export type components = { */ use_cache?: boolean; /** - * Start - * @description The start of the range - * @default 0 + * Tiles With Images + * @description A list of tile images with tile properties. + * @default null */ - start?: number; + tiles_with_images?: components["schemas"]["TileWithImage"][] | null; /** - * Size - * @description The number of values - * @default 1 + * Blend Mode + * @description blending type Linear or Seam + * @default Seam + * @enum {string} */ - size?: number; + blend_mode?: "Linear" | "Seam"; /** - * Step - * @description The step of the range - * @default 1 + * Blend Amount + * @description The amount to blend adjacent tiles in pixels. Must be <= the amount of overlap between adjacent tiles. + * @default 32 */ - step?: number; + blend_amount?: number; /** * type - * @default range_of_size + * @default merge_tiles_to_image * @constant */ - type: "range_of_size"; + type: "merge_tiles_to_image"; }; /** - * Reapply LoRA Weight - * @description Allows specifying an alternate weight for a LoRA after it has been selected with a LoRA selector node + * MetadataField + * @description Pydantic model for metadata with custom root of type dict[str, Any]. + * Metadata is stored without a strict schema. + */ + MetadataField: Record; + /** + * Metadata Field Extractor + * @description Extracts the text value from an image's metadata given a key. + * Raises an error if the image has no metadata or if the value is not a string (nesting not permitted). */ - ReapplyLoRAWeightInvocation: { + MetadataFieldExtractorInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -34811,199 +19636,229 @@ export type components = { */ use_cache?: boolean; /** - * LoRA + * @description The image to extract metadata from * @default null */ - lora?: components["schemas"]["LoRAField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 + * Key + * @description The key in the image's metadata to extract the value from + * @default null */ - weight?: number; + key?: string | null; /** * type - * @default reapply_lora_weight_invocation + * @default metadata_field_extractor * @constant */ - type: "reapply_lora_weight_invocation"; + type: "metadata_field_extractor"; }; /** - * ReapplyLoRAWeightOutput - * @description Reapply LoRA Weight Output + * Metadata From Image + * @description Used to create a core metadata item then Add/Update it to the provided metadata */ - ReapplyLoRAWeightOutput: { - /** - * LoRA - * @description LoRA model from a LoRA Selector node - */ - lora: components["schemas"]["LoRAField"]; + MetadataFromImageInvocation: { /** - * type - * @default reapply_lora_weight_output - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - type: "reapply_lora_weight_output"; - }; - /** - * RecallParameter - * @description Request model for updating recallable parameters. - */ - RecallParameter: { + id: string; /** - * Positive Prompt - * @description Positive prompt text + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - positive_prompt?: string | null; + is_intermediate?: boolean; /** - * Negative Prompt - * @description Negative prompt text + * Use Cache + * @description Whether or not to use the cache + * @default true */ - negative_prompt?: string | null; + use_cache?: boolean; /** - * Model - * @description Main model name/identifier + * @description The image to process + * @default null */ - model?: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Refiner Model - * @description Refiner model name/identifier + * type + * @default metadata_from_image + * @constant */ - refiner_model?: string | null; + type: "metadata_from_image"; + }; + /** + * Metadata + * @description Takes a MetadataItem or collection of MetadataItems and outputs a MetadataDict. + */ + MetadataInvocation: { /** - * Vae Model - * @description VAE model name/identifier + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - vae_model?: string | null; + id: string; /** - * Scheduler - * @description Scheduler name + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - scheduler?: string | null; + is_intermediate?: boolean; /** - * Steps - * @description Number of generation steps + * Use Cache + * @description Whether or not to use the cache + * @default true */ - steps?: number | null; + use_cache?: boolean; /** - * Refiner Steps - * @description Number of refiner steps + * Items + * @description A single metadata item or collection of metadata items + * @default null */ - refiner_steps?: number | null; + items?: components["schemas"]["MetadataItemField"][] | components["schemas"]["MetadataItemField"] | null; /** - * Cfg Scale - * @description CFG scale for guidance + * type + * @default metadata + * @constant */ - cfg_scale?: number | null; + type: "metadata"; + }; + /** MetadataItemField */ + MetadataItemField: { /** - * Cfg Rescale Multiplier - * @description CFG rescale multiplier + * Label + * @description Label for this metadata item */ - cfg_rescale_multiplier?: number | null; + label: string; /** - * Refiner Cfg Scale - * @description Refiner CFG scale + * Value + * @description The value for this metadata item (may be any type) */ - refiner_cfg_scale?: number | null; + value: unknown; + }; + /** + * Metadata Item + * @description Used to create an arbitrary metadata item. Provide "label" and make a connection to "value" to store that data as the value. + */ + MetadataItemInvocation: { /** - * Guidance - * @description Guidance scale + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - guidance?: number | null; + id: string; /** - * Width - * @description Image width in pixels + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - width?: number | null; + is_intermediate?: boolean; /** - * Height - * @description Image height in pixels + * Use Cache + * @description Whether or not to use the cache + * @default true */ - height?: number | null; + use_cache?: boolean; /** - * Seed - * @description Random seed + * Label + * @description Label for this metadata item + * @default null */ - seed?: number | null; + label?: string | null; /** - * Denoise Strength - * @description Denoising strength + * Value + * @description The value for this metadata item (may be any type) + * @default null */ - denoise_strength?: number | null; + value?: unknown | null; /** - * Refiner Denoise Start - * @description Refiner denoising start + * type + * @default metadata_item + * @constant */ - refiner_denoise_start?: number | null; + type: "metadata_item"; + }; + /** + * Metadata Item Linked + * @description Used to Create/Add/Update a value into a metadata label + */ + MetadataItemLinkedInvocation: { /** - * Clip Skip - * @description CLIP skip layers + * @description Optional metadata to be saved with the image + * @default null */ - clip_skip?: number | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Seamless X - * @description Enable seamless X tiling + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - seamless_x?: boolean | null; + id: string; /** - * Seamless Y - * @description Enable seamless Y tiling + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - seamless_y?: boolean | null; + is_intermediate?: boolean; /** - * Refiner Positive Aesthetic Score - * @description Refiner positive aesthetic score + * Use Cache + * @description Whether or not to use the cache + * @default true */ - refiner_positive_aesthetic_score?: number | null; + use_cache?: boolean; /** - * Refiner Negative Aesthetic Score - * @description Refiner negative aesthetic score + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - refiner_negative_aesthetic_score?: number | null; + label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt" | "width" | "height" | "seed" | "cfg_scale" | "cfg_rescale_multiplier" | "steps" | "scheduler" | "clip_skip" | "model" | "vae" | "seamless_x" | "seamless_y" | "guidance" | "cfg_scale_start_step" | "cfg_scale_end_step"; /** - * Loras - * @description List of LoRAs with their weights + * Custom Label + * @description Label for this metadata item + * @default null */ - loras?: components["schemas"]["LoRARecallParameter"][] | null; + custom_label?: string | null; /** - * Control Layers - * @description List of control adapters (ControlNet, T2I Adapter, Control LoRA) with their settings + * Value + * @description The value for this metadata item (may be any type) + * @default null */ - control_layers?: components["schemas"]["ControlNetRecallParameter"][] | null; + value?: unknown | null; /** - * Ip Adapters - * @description List of IP Adapters with their settings + * type + * @default metadata_item_linked + * @constant */ - ip_adapters?: components["schemas"]["IPAdapterRecallParameter"][] | null; + type: "metadata_item_linked"; }; /** - * RecallParametersUpdatedEvent - * @description Event model for recall_parameters_updated + * MetadataItemOutput + * @description Metadata Item Output */ - RecallParametersUpdatedEvent: { - /** - * Timestamp - * @description The timestamp of the event - */ - timestamp: number; + MetadataItemOutput: { + /** @description Metadata Item */ + item: components["schemas"]["MetadataItemField"]; /** - * Queue Id - * @description The ID of the queue + * type + * @default metadata_item_output + * @constant */ - queue_id: string; + type: "metadata_item_output"; + }; + /** MetadataOutput */ + MetadataOutput: { + /** @description Metadata Dict */ + metadata: components["schemas"]["MetadataField"]; /** - * Parameters - * @description The recall parameters that were updated + * type + * @default metadata_output + * @constant */ - parameters: { - [key: string]: unknown; - }; + type: "metadata_output"; }; /** - * Create Rectangle Mask - * @description Create a rectangular mask. + * Metadata To Bool Collection + * @description Extracts a Boolean value Collection of a label from metadata */ - RectangleMaskInvocation: { + MetadataToBoolCollectionInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35027,95 +19882,133 @@ export type components = { */ use_cache?: boolean; /** - * Width - * @description The width of the entire mask. - * @default null + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - width?: number | null; + label?: "* CUSTOM LABEL *" | "seamless_x" | "seamless_y"; /** - * Height - * @description The height of the entire mask. + * Custom Label + * @description Label for this metadata item * @default null */ - height?: number | null; + custom_label?: string | null; /** - * X Left - * @description The left x-coordinate of the rectangular masked region (inclusive). + * Default Value + * @description The default bool to use if not found in the metadata * @default null */ - x_left?: number | null; + default_value?: boolean[] | null; /** - * Y Top - * @description The top y-coordinate of the rectangular masked region (inclusive). + * type + * @default metadata_to_bool_collection + * @constant + */ + type: "metadata_to_bool_collection"; + }; + /** + * Metadata To Bool + * @description Extracts a Boolean value of a label from metadata + */ + MetadataToBoolInvocation: { + /** + * @description Optional metadata to be saved with the image * @default null */ - y_top?: number | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Rectangle Width - * @description The width of the rectangular masked region. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} + */ + label?: "* CUSTOM LABEL *" | "seamless_x" | "seamless_y"; + /** + * Custom Label + * @description Label for this metadata item * @default null */ - rectangle_width?: number | null; + custom_label?: string | null; /** - * Rectangle Height - * @description The height of the rectangular masked region. + * Default Value + * @description The default bool to use if not found in the metadata * @default null */ - rectangle_height?: number | null; + default_value?: boolean | null; /** * type - * @default rectangle_mask + * @default metadata_to_bool * @constant */ - type: "rectangle_mask"; + type: "metadata_to_bool"; }; /** - * RemoteModelFile - * @description Information about a downloadable file that forms part of a model. + * Metadata To ControlNets + * @description Extracts a Controlnets value of a label from metadata */ - RemoteModelFile: { + MetadataToControlnetsInvocation: { /** - * Url - * Format: uri - * @description The url to download this model file + * @description Optional metadata to be saved with the image + * @default null */ - url: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Path - * Format: path - * @description The path to the file, relative to the model root + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * Size - * @description The size of this file, in bytes - * @default 0 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - size?: number | null; + is_intermediate?: boolean; /** - * Sha256 - * @description SHA256 hash of this model (not always available) + * Use Cache + * @description Whether or not to use the cache + * @default true */ - sha256?: string | null; - }; - /** RemoveImagesFromBoardResult */ - RemoveImagesFromBoardResult: { + use_cache?: boolean; /** - * Affected Boards - * @description The ids of boards affected by the delete operation + * ControlNet-List + * @default null */ - affected_boards: string[]; + control_list?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; /** - * Removed Images - * @description The image names that were removed from their board + * type + * @default metadata_to_controlnets + * @constant */ - removed_images: string[]; + type: "metadata_to_controlnets"; }; /** - * Resize Latents - * @description Resizes latents to explicit width/height (in pixels). Provided dimensions are floor-divided by 8. + * Metadata To Float Collection + * @description Extracts a Float value Collection of a label from metadata */ - ResizeLatentsInvocation: { + MetadataToFloatCollectionInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -35134,54 +20027,36 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - latents?: components["schemas"]["LatentsField"] | null; + label?: "* CUSTOM LABEL *" | "cfg_scale" | "cfg_rescale_multiplier" | "guidance"; /** - * Width - * @description Width of output (px) + * Custom Label + * @description Label for this metadata item * @default null */ - width?: number | null; + custom_label?: string | null; /** - * Height - * @description Width of output (px) + * Default Value + * @description The default float to use if not found in the metadata * @default null */ - height?: number | null; - /** - * Mode - * @description Interpolation mode - * @default bilinear - * @enum {string} - */ - mode?: "nearest" | "linear" | "bilinear" | "bicubic" | "trilinear" | "area" | "nearest-exact"; - /** - * Antialias - * @description Whether or not to apply antialiasing (bilinear or bicubic only) - * @default false - */ - antialias?: boolean; + default_value?: number[] | null; /** * type - * @default lresize + * @default metadata_to_float_collection * @constant */ - type: "lresize"; + type: "metadata_to_float_collection"; }; /** - * ResourceOrigin - * @description The origin of a resource (eg image). - * - * - INTERNAL: The resource was created by the application. - * - EXTERNAL: The resource was not created by the application. - * This may be a user-initiated upload, or an internal application upload (eg Canvas init image). - * @enum {string} + * Metadata To Float + * @description Extracts a Float value of a label from metadata */ - ResourceOrigin: "internal" | "external"; - /** Retrieve Images from Board */ - RetrieveBoardImagesInvocation: { + MetadataToFloatInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35201,55 +20076,45 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** - * @description Input board containing images to be retrieved. If not provided or set to None, it will grab images from uncategorized. - * @default null - */ - input_board?: components["schemas"]["BoardField"] | null; - /** - * Num Images - * @description Number of images to retrieve: specify 'all', a single index like '10', specific indices like '1,4,6', or a range like '30-50'. - * @default all - */ - num_images?: string; - /** - * Category - * @description Category of images to retrieve; select either 'images' or 'assets'. - * @default images + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * * @enum {string} */ - category?: "images" | "assets"; + label?: "* CUSTOM LABEL *" | "cfg_scale" | "cfg_rescale_multiplier" | "guidance"; /** - * Starred Only - * @description Retrieve only starred images if set to True. - * @default false + * Custom Label + * @description Label for this metadata item + * @default null */ - starred_only?: boolean; + custom_label?: string | null; /** - * Keyword - * @description Keyword to filter images by metadata. Only images with metadata containing this keyword will be retrieved. + * Default Value + * @description The default float to use if not found in the metadata * @default null */ - keyword?: string | null; + default_value?: number | null; /** * type - * @default Retrieve_Board_Images + * @default metadata_to_float * @constant */ - type: "Retrieve_Board_Images"; + type: "metadata_to_float"; }; /** - * Retrieve Flux Conditioning - * @description Retrieves one or more FLUX Conditioning objects (CLIP and T5 embeddings) - * from an SQLite database using unique identifiers. - * Outputs a single selected conditioning and a list of all retrieved conditionings. - * Includes an option to update the timestamp of retrieved entries. - * The input conditioning IDs are sorted for deterministic retrieval. + * Metadata To IP-Adapters + * @description Extracts a IP-Adapters value of a label from metadata */ - RetrieveFluxConditioningInvocation: { + MetadataToIPAdaptersInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -35268,55 +20133,23 @@ export type components = { */ use_cache?: boolean; /** - * Conditioning Id Or List - * @description The unique identifier(s) of the Flux Conditioning(s) to retrieve. + * IP-Adapter-List + * @description IP-Adapter to apply * @default null */ - conditioning_id_or_list?: string | string[] | null; - /** - * Select Index - * @description Index of the retrieved conditioning to output as the single 'conditioning' field. If out of bounds, uses modulus. - * @default 0 - */ - select_index?: number; - /** - * Touch Timestamp - * @description When true, updates the timestamp of retrieved entries to 'now', preventing early purge. - * @default false - */ - touch_timestamp?: boolean; - /** - * type - * @default retrieve_flux_conditioning - * @constant - */ - type: "retrieve_flux_conditioning"; - }; - /** - * RetrieveFluxConditioningMultiOutput - * @description Output for the Retrieve Flux Conditioning node, providing a single selected - * conditioning and a list of all retrieved conditionings. - */ - RetrieveFluxConditioningMultiOutput: { - /** @description A single selected Flux Conditioning (selected by index from the retrieved list) */ - conditioning: components["schemas"]["FluxConditioningField"]; - /** - * Conditioning List - * @description A list of all retrieved Flux Conditionings - */ - conditioning_list: components["schemas"]["FluxConditioningField"][]; + ip_adapter_list?: components["schemas"]["IPAdapterField"] | components["schemas"]["IPAdapterField"][] | null; /** * type - * @default retrieve_flux_conditioning_multi_output + * @default metadata_to_ip_adapters * @constant */ - type: "retrieve_flux_conditioning_multi_output"; + type: "metadata_to_ip_adapters"; }; /** - * Bitize - * @description Crush an image to one-bit pixels + * Metadata To Integer Collection + * @description Extracts an integer value Collection of a label from metadata */ - RetroBitizeInvocation: { + MetadataToIntegerCollectionInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35340,28 +20173,36 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image for pixelization + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} + */ + label?: "* CUSTOM LABEL *" | "width" | "height" | "seed" | "steps" | "clip_skip" | "cfg_scale_start_step" | "cfg_scale_end_step"; + /** + * Custom Label + * @description Label for this metadata item * @default null */ - image?: components["schemas"]["ImageField"] | null; + custom_label?: string | null; /** - * Dither - * @description Dither the Bitized image - * @default true + * Default Value + * @description The default integer to use if not found in the metadata + * @default null */ - dither?: boolean; + default_value?: number[] | null; /** * type - * @default retro_bitize + * @default metadata_to_integer_collection * @constant */ - type: "retro_bitize"; + type: "metadata_to_integer_collection"; }; /** - * CRT - * @description Distort the input image, simulating CRT display curvature + * Metadata To Integer + * @description Extracts an integer value of a label from metadata */ - RetroCRTCurvatureInvocation: { + MetadataToIntegerInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35385,64 +20226,36 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image for pixelization - * @default null - */ - image?: components["schemas"]["ImageField"] | null; - /** - * Crt Width - * @description Horizontal resolution of the CRT; smaller = more noticeable - * @default 240 - */ - crt_width?: number; - /** - * Crt Height - * @description Vertical resolution of the CRT; smaller = more noticeable - * @default 160 - */ - crt_height?: number; - /** - * Crt Curvature - * @description Curvature factor; smaller = stronger inward curve - * @default 3 - */ - crt_curvature?: number; - /** - * Scanlines Opacity - * @description Opacity of CRT scan lines - * @default 1 - */ - scanlines_opacity?: number; - /** - * Vignette Opacity - * @description Vignette opacity - * @default 0.5 + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - vignette_opacity?: number; + label?: "* CUSTOM LABEL *" | "width" | "height" | "seed" | "steps" | "clip_skip" | "cfg_scale_start_step" | "cfg_scale_end_step"; /** - * Vignette Roundness - * @description Vignette opacity - * @default 5 + * Custom Label + * @description Label for this metadata item + * @default null */ - vignette_roundness?: number; + custom_label?: string | null; /** - * Crt Brightness - * @description Factor by which to brighten the image if using scan lines - * @default 1.2 + * Default Value + * @description The default integer to use if not found in the metadata + * @default null */ - crt_brightness?: number; + default_value?: number | null; /** * type - * @default retro_crt_curvature + * @default metadata_to_integer * @constant */ - type: "retro_crt_curvature"; + type: "metadata_to_integer"; }; /** - * Get Palette (Advanced) - * @description Get palette from an image, 256 colors max. Optionally export to a user-defined location. + * Metadata To LoRA Collection + * @description Extracts Lora(s) from metadata into a collection */ - RetroGetPaletteAdvInvocation: { + MetadataToLorasCollectionInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35466,40 +20279,46 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image to grab a palette from - * @default null + * Custom Label + * @description Label for this metadata item + * @default loras */ - image?: components["schemas"]["ImageField"] | null; + custom_label?: string; /** - * Export - * @description Save palette PNG to specified path with optional name - * @default true + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. + * @default [] */ - export?: boolean; + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * Subfolder - * @description Subfolder for the palette in nodes/Retroize/palettes/ folder - * @default + * type + * @default metadata_to_lora_collection + * @constant */ - subfolder?: string; + type: "metadata_to_lora_collection"; + }; + /** + * MetadataToLorasCollectionOutput + * @description Model loader output + */ + MetadataToLorasCollectionOutput: { /** - * Name - * @description Name for the palette image - * @default + * LoRAs + * @description Collection of LoRA model and weights */ - name?: string; + lora: components["schemas"]["LoRAField"][]; /** * type - * @default get_palette_adv + * @default metadata_to_lora_collection_output * @constant */ - type: "get_palette_adv"; + type: "metadata_to_lora_collection_output"; }; /** - * Get Palette - * @description Get palette from an image, 256 colors max. + * Metadata To LoRAs + * @description Extracts a Loras value of a label from metadata */ - RetroGetPaletteInvocation: { + MetadataToLorasInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35523,22 +20342,29 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image to grab a palette from + * UNet + * @description UNet (scheduler, LoRAs) * @default null */ - image?: components["schemas"]["ImageField"] | null; + unet?: components["schemas"]["UNetField"] | null; + /** + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; /** * type - * @default get_palette + * @default metadata_to_loras * @constant */ - type: "get_palette"; + type: "metadata_to_loras"; }; /** - * Palettize Advanced - * @description Palettize an image by applying a color palette + * Metadata To Model + * @description Extracts a Model value of a label from metadata */ - RetroPalettizeAdvInvocation: { + MetadataToModelInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35562,116 +20388,72 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image for pixelization - * @default null + * Label + * @description Label for this metadata item + * @default model + * @enum {string} */ - image?: components["schemas"]["ImageField"] | null; + label?: "* CUSTOM LABEL *" | "model"; /** - * Palette Image - * @description Palette image + * Custom Label + * @description Label for this metadata item * @default null */ - palette_image?: ("None" | "aap-64.png" | "atari-8-bit.png" | "commodore64.png" | "endesga-32.png" | "fantasy-24.png" | "mg\\mg-16-01.png" | "microsoft-windows.png" | "NES.png" | "nintendo-gameboy.png" | "pico-8.png" | "slso8.png") | null; - /** - * Dither - * @description Apply dithering to image when palettizing - * @default false - */ - dither?: boolean; - /** - * Prequantize - * @description Apply 256-color quantization with specified method prior to applying the color palette - * @default false - */ - prequantize?: boolean; + custom_label?: string | null; /** - * Quantizer - * @description Palettizer quantization method - * @default Fast Octree - * @enum {string} + * @description The default model to use if not found in the metadata + * @default null */ - quantizer?: "Median Cut" | "Max Coverage" | "Fast Octree"; + default_value?: components["schemas"]["ModelIdentifierField"] | null; /** * type - * @default retro_palettize_adv + * @default metadata_to_model * @constant */ - type: "retro_palettize_adv"; + type: "metadata_to_model"; }; /** - * Palettize - * @description Palettize an image by applying a color palette + * MetadataToModelOutput + * @description String to main model output */ - RetroPalettizeInvocation: { - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** - * @description Input image for pixelization - * @default null - */ - image?: components["schemas"]["ImageField"] | null; + MetadataToModelOutput: { /** - * @description Palette image - * @default null + * Model + * @description Main model (UNet, VAE, CLIP) to load */ - palette_image?: components["schemas"]["ImageField"] | null; + model: components["schemas"]["ModelIdentifierField"]; /** - * Palette Path - * @description Palette image path, including ".png" extension - * @default + * Name + * @description Model Name */ - palette_path?: string; + name: string; /** - * Dither - * @description Apply dithering to image when palettizing - * @default false + * UNet + * @description UNet (scheduler, LoRAs) */ - dither?: boolean; + unet: components["schemas"]["UNetField"]; /** - * Prequantize - * @description Apply 256-color quantization with specified method prior to applying the color palette - * @default false + * VAE + * @description VAE */ - prequantize?: boolean; + vae: components["schemas"]["VAEField"]; /** - * Quantizer - * @description Palettizer quantization method - * @default Fast Octree - * @enum {string} + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - quantizer?: "Median Cut" | "Max Coverage" | "Fast Octree"; + clip: components["schemas"]["CLIPField"]; /** * type - * @default retro_palettize + * @default metadata_to_model_output * @constant */ - type: "retro_palettize"; + type: "metadata_to_model_output"; }; /** - * Quantize - * @description Quantize an image to 256 or less colors + * Metadata To SDXL LoRAs + * @description Extracts a SDXL Loras value of a label from metadata */ - RetroQuantizeInvocation: { + MetadataToSDXLLorasInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35695,47 +20477,35 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image for quantizing + * UNet + * @description UNet (scheduler, LoRAs) * @default null */ - image?: components["schemas"]["ImageField"] | null; - /** - * Colors - * @description Number of colors the image should be reduced to - * @default 64 - */ - colors?: number; - /** - * Method - * @description Quantization method - * @default Median Cut - * @enum {string} - */ - method?: "Median Cut" | "Max Coverage" | "Fast Octree"; + unet?: components["schemas"]["UNetField"] | null; /** - * Kmeans - * @description k_means - * @default 0 + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - kmeans?: number; + clip?: components["schemas"]["CLIPField"] | null; /** - * Dither - * @description Dither quantized image - * @default true + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - dither?: boolean; + clip2?: components["schemas"]["CLIPField"] | null; /** * type - * @default retro_quantize + * @default metadata_to_sdlx_loras * @constant */ - type: "retro_quantize"; + type: "metadata_to_sdlx_loras"; }; /** - * Scan Lines - * @description Apply a simple scan lines effect to the input image + * Metadata To SDXL Model + * @description Extracts a SDXL Model value of a label from metadata */ - RetroScanlinesSimpleInvocation: { + MetadataToSDXLModelInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35759,75 +20529,82 @@ export type components = { */ use_cache?: boolean; /** - * @description Input image to add scanlines to + * Label + * @description Label for this metadata item + * @default model + * @enum {string} + */ + label?: "* CUSTOM LABEL *" | "model"; + /** + * Custom Label + * @description Label for this metadata item * @default null */ - image?: components["schemas"]["ImageField"] | null; + custom_label?: string | null; /** - * Line Size - * @description Thickness of scanlines in pixels - * @default 1 + * @description The default SDXL Model to use if not found in the metadata + * @default null */ - line_size?: number; + default_value?: components["schemas"]["ModelIdentifierField"] | null; /** - * Line Spacing - * @description Space between lines in pixels - * @default 4 + * type + * @default metadata_to_sdxl_model + * @constant */ - line_spacing?: number; + type: "metadata_to_sdxl_model"; + }; + /** + * MetadataToSDXLModelOutput + * @description String to SDXL main model output + */ + MetadataToSDXLModelOutput: { /** - * @description Darkness of scanlines, 1.0 being black - * @default { - * "r": 0, - * "g": 0, - * "b": 0, - * "a": 255 - * } + * Model + * @description Main model (UNet, VAE, CLIP) to load */ - line_color?: components["schemas"]["ColorField"]; + model: components["schemas"]["ModelIdentifierField"]; /** - * Size Jitter - * @description Random line position offset - * @default 0 + * Name + * @description Model Name */ - size_jitter?: number; + name: string; /** - * Space Jitter - * @description Random line position offset - * @default 0 + * UNet + * @description UNet (scheduler, LoRAs) */ - space_jitter?: number; + unet: components["schemas"]["UNetField"]; /** - * Vertical - * @description Switch scanlines to vertical - * @default false + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - vertical?: boolean; + clip: components["schemas"]["CLIPField"]; /** - * type - * @default retro_scanlines_simple - * @constant + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - type: "retro_scanlines_simple"; - }; - /** RetryItemsResult */ - RetryItemsResult: { + clip2: components["schemas"]["CLIPField"]; /** - * Queue Id - * @description The ID of the queue + * VAE + * @description VAE */ - queue_id: string; + vae: components["schemas"]["VAEField"]; /** - * Retried Item Ids - * @description The IDs of the queue items that were retried + * type + * @default metadata_to_sdxl_model_output + * @constant */ - retried_item_ids: number[]; + type: "metadata_to_sdxl_model_output"; }; /** - * Round Float - * @description Rounds a float to a specified number of decimal places. + * Metadata To Scheduler + * @description Extracts a Scheduler value of a label from metadata */ - RoundInvocation: { + MetadataToSchedulerInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -35846,87 +20623,37 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The float value - * @default 0 - */ - value?: number; - /** - * Decimals - * @description The number of decimal places - * @default 0 - */ - decimals?: number; - /** - * type - * @default round_float - * @constant - */ - type: "round_float"; - }; - /** SAMPoint */ - SAMPoint: { - /** - * X - * @description The x-coordinate of the point - */ - x: number; - /** - * Y - * @description The y-coordinate of the point + * Label + * @description Label for this metadata item + * @default scheduler + * @enum {string} */ - y: number; - /** @description The label of the point */ - label: components["schemas"]["SAMPointLabel"]; - }; - /** - * SAMPointLabel - * @enum {integer} - */ - SAMPointLabel: -1 | 0 | 1; - /** SAMPointsField */ - SAMPointsField: { + label?: "* CUSTOM LABEL *" | "scheduler"; /** - * Points - * @description The points of the object - */ - points: components["schemas"]["SAMPoint"][]; - }; - /** - * SD3ConditioningField - * @description A conditioning tensor primitive value - */ - SD3ConditioningField: { + * Custom Label + * @description Label for this metadata item + * @default null + */ + custom_label?: string | null; /** - * Conditioning Name - * @description The name of conditioning tensor + * Default Value + * @description The default scheduler to use if not found in the metadata + * @default euler + * @enum {string} */ - conditioning_name: string; - }; - /** - * SD3ConditioningOutput - * @description Base class for nodes that output a single SD3 conditioning tensor - */ - SD3ConditioningOutput: { - /** @description Conditioning tensor */ - conditioning: components["schemas"]["SD3ConditioningField"]; + default_value?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** * type - * @default sd3_conditioning_output + * @default metadata_to_scheduler * @constant */ - type: "sd3_conditioning_output"; + type: "metadata_to_scheduler"; }; /** - * Denoise - SD3 - * @description Run denoising process with a SD3 model. + * Metadata To String Collection + * @description Extracts a string collection value of a label from metadata */ - SD3DenoiseInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + MetadataToStringCollectionInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -35950,90 +20677,129 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor - * @default null + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} */ - latents?: components["schemas"]["LatentsField"] | null; + label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt"; /** - * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. + * Custom Label + * @description Label for this metadata item * @default null */ - denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + custom_label?: string | null; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Default Value + * @description The default string collection to use if not found in the metadata + * @default null */ - denoising_start?: number; + default_value?: string[] | null; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * type + * @default metadata_to_string_collection + * @constant */ - denoising_end?: number; + type: "metadata_to_string_collection"; + }; + /** + * Metadata To String + * @description Extracts a string value of a label from metadata + */ + MetadataToStringInvocation: { /** - * Transformer - * @description SD3 model (MMDiTX) to load + * @description Optional metadata to be saved with the image * @default null */ - transformer?: components["schemas"]["TransformerField"] | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * @description Positive conditioning tensor + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * Label + * @description Label for this metadata item + * @default * CUSTOM LABEL * + * @enum {string} + */ + label?: "* CUSTOM LABEL *" | "positive_prompt" | "positive_style_prompt" | "negative_prompt" | "negative_style_prompt"; + /** + * Custom Label + * @description Label for this metadata item * @default null */ - positive_conditioning?: components["schemas"]["SD3ConditioningField"] | null; + custom_label?: string | null; /** - * @description Negative conditioning tensor + * Default Value + * @description The default string to use if not found in the metadata * @default null */ - negative_conditioning?: components["schemas"]["SD3ConditioningField"] | null; + default_value?: string | null; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 3.5 + * type + * @default metadata_to_string + * @constant */ - cfg_scale?: number | number[]; + type: "metadata_to_string"; + }; + /** + * Metadata To T2I-Adapters + * @description Extracts a T2I-Adapters value of a label from metadata + */ + MetadataToT2IAdaptersInvocation: { /** - * Width - * @description Width of the generated image. - * @default 1024 + * @description Optional metadata to be saved with the image + * @default null */ - width?: number; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Height - * @description Height of the generated image. - * @default 1024 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - height?: number; + id: string; /** - * Steps - * @description Number of steps to run - * @default 10 + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - steps?: number; + is_intermediate?: boolean; /** - * Seed - * @description Randomness seed for reproducibility. - * @default 0 + * Use Cache + * @description Whether or not to use the cache + * @default true */ - seed?: number; + use_cache?: boolean; + /** + * T2I-Adapter + * @description IP-Adapter to apply + * @default null + */ + t2i_adapter_list?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; /** * type - * @default sd3_denoise + * @default metadata_to_t2i_adapters * @constant */ - type: "sd3_denoise"; + type: "metadata_to_t2i_adapters"; }; /** - * Image to Latents - SD3 - * @description Generates latents from an image. + * Metadata To VAE + * @description Extracts a VAE value of a label from metadata */ - SD3ImageToLatentsInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; + MetadataToVAEInvocation: { /** * @description Optional metadata to be saved with the image * @default null @@ -36057,37 +20823,70 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to encode + * Label + * @description Label for this metadata item + * @default vae + * @enum {string} + */ + label?: "* CUSTOM LABEL *" | "vae"; + /** + * Custom Label + * @description Label for this metadata item * @default null */ - image?: components["schemas"]["ImageField"] | null; + custom_label?: string | null; /** - * @description VAE + * @description The default VAE to use if not found in the metadata * @default null */ - vae?: components["schemas"]["VAEField"] | null; + default_value?: components["schemas"]["VAEField"] | null; /** * type - * @default sd3_i2l + * @default metadata_to_vae * @constant */ - type: "sd3_i2l"; + type: "metadata_to_vae"; }; /** - * Latents to Image - SD3 - * @description Generates an image from latents. + * ModelFormat + * @description Storage format of model. + * @enum {string} */ - SD3LatentsToImageInvocation: { + ModelFormat: "omi" | "diffusers" | "checkpoint" | "lycoris" | "onnx" | "olive" | "embedding_file" | "embedding_folder" | "invokeai" | "t5_encoder" | "qwen3_encoder" | "bnb_quantized_int8b" | "bnb_quantized_nf4b" | "gguf_quantized" | "unknown"; + /** ModelIdentifierField */ + ModelIdentifierField: { /** - * @description The board to save the image to - * @default null + * Key + * @description The model's unique key */ - board?: components["schemas"]["BoardField"] | null; + key: string; /** - * @description Optional metadata to be saved with the image + * Hash + * @description The model's BLAKE3 hash + */ + hash: string; + /** + * Name + * @description The model's name + */ + name: string; + /** @description The model's base model type */ + base: components["schemas"]["BaseModelType"]; + /** @description The model's type */ + type: components["schemas"]["ModelType"]; + /** + * @description The submodel to load, if this is a main model * @default null */ - metadata?: components["schemas"]["MetadataField"] | null; + submodel_type?: components["schemas"]["SubModelType"] | null; + }; + /** + * Any Model + * @description Selects any model, outputting it its identifier. Be careful with this one! The identifier will be accepted as + * input for any model, even if the model types don't match. If you connect this to a mismatched input, you'll get an + * error. + */ + ModelIdentifierInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -36106,532 +20905,550 @@ export type components = { */ use_cache?: boolean; /** - * @description Latents tensor + * Model + * @description The model to select * @default null */ - latents?: components["schemas"]["LatentsField"] | null; + model?: components["schemas"]["ModelIdentifierField"] | null; /** - * @description VAE - * @default null + * type + * @default model_identifier + * @constant + */ + type: "model_identifier"; + }; + /** + * ModelIdentifierOutput + * @description Model identifier output + */ + ModelIdentifierOutput: { + /** + * Model + * @description Model identifier */ - vae?: components["schemas"]["VAEField"] | null; + model: components["schemas"]["ModelIdentifierField"]; /** * type - * @default sd3_l2i + * @default model_identifier_output * @constant */ - type: "sd3_l2i"; + type: "model_identifier_output"; }; /** - * SD3 Main Model Input - * @description Loads a sd3 model from an input, outputting its submodels. + * ModelInstallCancelledEvent + * @description Event model for model_install_cancelled */ - SD3ModelLoaderInputInvocation: { + ModelInstallCancelledEvent: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Timestamp + * @description The timestamp of the event */ - id: string; + timestamp: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Id + * @description The ID of the install job */ - is_intermediate?: boolean; + id: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source + * @description Source of the model; local path, repo_id or url */ - use_cache?: boolean; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + }; + /** + * ModelInstallCompleteEvent + * @description Event model for model_install_complete + */ + ModelInstallCompleteEvent: { /** - * @description SD3 model (MMDiTX) to load - * @default null + * Timestamp + * @description The timestamp of the event */ - model?: components["schemas"]["ModelIdentifierField"] | null; + timestamp: number; /** - * T5 Encoder - * @description T5 tokenizer and text encoder - * @default null + * Id + * @description The ID of the install job */ - t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; + id: number; /** - * CLIP L Encoder - * @description CLIP Embed loader - * @default null + * Source + * @description Source of the model; local path, repo_id or url */ - clip_l_model?: components["schemas"]["ModelIdentifierField"] | null; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * CLIP G Encoder - * @description CLIP-G Embed loader - * @default null + * Key + * @description Model config record key */ - clip_g_model?: components["schemas"]["ModelIdentifierField"] | null; + key: string; /** - * VAE - * @description VAE model to load - * @default null + * Total Bytes + * @description Size of the model (may be None for installation of a local path) */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + total_bytes: number | null; /** - * type - * @default sd3_model_loader_input - * @constant + * Config + * @description The installed model's config */ - type: "sd3_model_loader_input"; + config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; }; /** - * Prompt - SDXL - * @description Parse prompt using compel package to conditioning. + * ModelInstallDownloadProgressEvent + * @description Event model for model_install_download_progress */ - SDXLCompelPromptInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + ModelInstallDownloadProgressEvent: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Timestamp + * @description The timestamp of the event */ - is_intermediate?: boolean; + timestamp: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Id + * @description The ID of the install job */ - use_cache?: boolean; + id: number; /** - * Prompt - * @description Prompt to be parsed by Compel to create a conditioning tensor - * @default + * Source + * @description Source of the model; local path, repo_id or url */ - prompt?: string; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * Style - * @description Prompt to be parsed by Compel to create a conditioning tensor - * @default + * Local Path + * @description Where model is downloading to */ - style?: string; + local_path: string; /** - * Original Width - * @default 1024 + * Bytes + * @description Number of bytes downloaded so far */ - original_width?: number; + bytes: number; /** - * Original Height - * @default 1024 + * Total Bytes + * @description Total size of download, including all files */ - original_height?: number; + total_bytes: number; /** - * Crop Top - * @default 0 + * Parts + * @description Progress of downloading URLs that comprise the model, if any */ - crop_top?: number; + parts: { + [key: string]: number | string; + }[]; + }; + /** + * ModelInstallDownloadStartedEvent + * @description Event model for model_install_download_started + */ + ModelInstallDownloadStartedEvent: { /** - * Crop Left - * @default 0 + * Timestamp + * @description The timestamp of the event */ - crop_left?: number; + timestamp: number; /** - * Target Width - * @default 1024 + * Id + * @description The ID of the install job */ - target_width?: number; + id: number; /** - * Target Height - * @default 1024 + * Source + * @description Source of the model; local path, repo_id or url */ - target_height?: number; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Local Path + * @description Where model is downloading to */ - clip?: components["schemas"]["CLIPField"] | null; + local_path: string; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Bytes + * @description Number of bytes downloaded so far */ - clip2?: components["schemas"]["CLIPField"] | null; + bytes: number; /** - * @description A mask defining the region that this conditioning prompt applies to. - * @default null + * Total Bytes + * @description Total size of download, including all files */ - mask?: components["schemas"]["TensorField"] | null; + total_bytes: number; /** - * type - * @default sdxl_compel_prompt - * @constant + * Parts + * @description Progress of downloading URLs that comprise the model, if any */ - type: "sdxl_compel_prompt"; + parts: { + [key: string]: number | string; + }[]; }; /** - * Apply LoRA Collection - SDXL - * @description Applies a collection of SDXL LoRAs to the provided UNet and CLIP models. + * ModelInstallDownloadsCompleteEvent + * @description Emitted once when an install job becomes active. */ - SDXLLoRACollectionLoader: { + ModelInstallDownloadsCompleteEvent: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Timestamp + * @description The timestamp of the event */ - id: string; + timestamp: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Id + * @description The ID of the install job */ - is_intermediate?: boolean; + id: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source + * @description Source of the model; local path, repo_id or url */ - use_cache?: boolean; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + }; + /** + * ModelInstallErrorEvent + * @description Event model for model_install_error + */ + ModelInstallErrorEvent: { /** - * LoRAs - * @description LoRA models and weights. May be a single LoRA or collection. - * @default null + * Timestamp + * @description The timestamp of the event */ - loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; + timestamp: number; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Id + * @description The ID of the install job */ - unet?: components["schemas"]["UNetField"] | null; + id: number; /** - * CLIP - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Source + * @description Source of the model; local path, repo_id or url */ - clip?: components["schemas"]["CLIPField"] | null; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Error Type + * @description The name of the exception */ - clip2?: components["schemas"]["CLIPField"] | null; + error_type: string; /** - * type - * @default sdxl_lora_collection_loader - * @constant + * Error + * @description A text description of the exception */ - type: "sdxl_lora_collection_loader"; + error: string; }; /** - * Apply LoRA - SDXL - * @description Apply selected lora to unet and text_encoder. + * ModelInstallJob + * @description Object that tracks the current status of an install request. */ - SDXLLoRALoaderInvocation: { + ModelInstallJob: { /** * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description Unique ID for this job */ - id: string; + id: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * @description Current status of install process + * @default waiting */ - is_intermediate?: boolean; + status?: components["schemas"]["InstallStatus"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Error Reason + * @description Information about why the job failed */ - use_cache?: boolean; + error_reason?: string | null; + /** @description Configuration information (e.g. 'description') to apply to model. */ + config_in?: components["schemas"]["ModelRecordChanges"]; /** - * LoRA - * @description LoRA model to load - * @default null + * Config Out + * @description After successful installation, this will hold the configuration object. */ - lora?: components["schemas"]["ModelIdentifierField"] | null; + config_out?: (components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]) | null; /** - * Weight - * @description The weight at which the LoRA is applied to each model - * @default 0.75 + * Inplace + * @description Leave model in its current location; otherwise install under models directory + * @default false */ - weight?: number; + inplace?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Source + * @description Source (URL, repo_id, or local path) of model */ - unet?: components["schemas"]["UNetField"] | null; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Local Path + * Format: path + * @description Path to locally-downloaded model; may be the same as the source */ - clip?: components["schemas"]["CLIPField"] | null; + local_path: string; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Bytes + * @description For a remote model, the number of bytes downloaded so far (may not be available) + * @default 0 */ - clip2?: components["schemas"]["CLIPField"] | null; + bytes?: number; /** - * type - * @default sdxl_lora_loader - * @constant + * Total Bytes + * @description Total size of the model to be installed + * @default 0 */ - type: "sdxl_lora_loader"; - }; - /** - * SDXLLoRALoaderOutput - * @description SDXL LoRA Loader Output - */ - SDXLLoRALoaderOutput: { + total_bytes?: number; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Source Metadata + * @description Metadata provided by the model source */ - unet: components["schemas"]["UNetField"] | null; + source_metadata?: (components["schemas"]["BaseMetadata"] | components["schemas"]["HuggingFaceMetadata"]) | null; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Download Parts + * @description Download jobs contributing to this install */ - clip: components["schemas"]["CLIPField"] | null; + download_parts?: components["schemas"]["DownloadJob"][]; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Error + * @description On an error condition, this field will contain the text of the exception */ - clip2: components["schemas"]["CLIPField"] | null; + error?: string | null; /** - * type - * @default sdxl_lora_loader_output - * @constant + * Error Traceback + * @description On an error condition, this field will contain the exception traceback */ - type: "sdxl_lora_loader_output"; + error_traceback?: string | null; }; /** - * SDXL Main Model Toggle - * @description Allows boolean selection between two separate SDXL Main Model inputs + * ModelInstallStartedEvent + * @description Event model for model_install_started */ - SDXLMainModelToggleInvocation: { + ModelInstallStartedEvent: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Timestamp + * @description The timestamp of the event */ - id: string; + timestamp: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Id + * @description The ID of the install job */ - is_intermediate?: boolean; + id: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source + * @description Source of the model; local path, repo_id or url */ - use_cache?: boolean; + source: components["schemas"]["LocalModelSource"] | components["schemas"]["HFModelSource"] | components["schemas"]["URLModelSource"]; + }; + /** + * ModelLoadCompleteEvent + * @description Event model for model_load_complete + */ + ModelLoadCompleteEvent: { /** - * Use Second - * @description Use 2nd Input - * @default false + * Timestamp + * @description The timestamp of the event */ - use_second?: boolean; + timestamp: number; /** - * UNet 1 - * @description UNet (scheduler, LoRAs) - * @default null + * Config + * @description The model's config */ - unet1?: components["schemas"]["UNetField"] | null; + config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; /** - * CLIP 1 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @description The submodel type, if any * @default null */ - clip1?: components["schemas"]["CLIPField"] | null; + submodel_type: components["schemas"]["SubModelType"] | null; + }; + /** + * ModelLoadStartedEvent + * @description Event model for model_load_started + */ + ModelLoadStartedEvent: { /** - * CLIP 2 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Timestamp + * @description The timestamp of the event */ - clip21?: components["schemas"]["CLIPField"] | null; + timestamp: number; /** - * VAE 1 - * @description VAE - * @default null + * Config + * @description The model's config */ - vae1?: components["schemas"]["VAEField"] | null; + config: components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"]; /** - * UNet 2 - * @description UNet (scheduler, LoRAs) + * @description The submodel type, if any * @default null */ - unet2?: components["schemas"]["UNetField"] | null; + submodel_type: components["schemas"]["SubModelType"] | null; + }; + /** + * ModelLoaderOutput + * @description Model loader output + */ + ModelLoaderOutput: { /** - * CLIP 1 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * VAE + * @description VAE */ - clip2?: components["schemas"]["CLIPField"] | null; + vae: components["schemas"]["VAEField"]; /** - * CLIP 2 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * type + * @default model_loader_output + * @constant */ - clip22?: components["schemas"]["CLIPField"] | null; + type: "model_loader_output"; /** - * VAE 2 - * @description VAE - * @default null + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - vae2?: components["schemas"]["VAEField"] | null; + clip: components["schemas"]["CLIPField"]; /** - * type - * @default sdxl_main_model_toggle - * @constant + * UNet + * @description UNet (scheduler, LoRAs) */ - type: "sdxl_main_model_toggle"; + unet: components["schemas"]["UNetField"]; }; /** - * SDXL Main Model Input - * @description Loads a sdxl model from an input, outputting its submodels. + * ModelRecordChanges + * @description A set of changes to apply to a model. */ - SDXLModelLoaderInputInvocation: { + ModelRecordChanges: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source + * @description original source of the model */ - id: string; + source?: string | null; + /** @description type of model source */ + source_type?: components["schemas"]["ModelSourceType"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source Api Response + * @description metadata from remote source */ - is_intermediate?: boolean; + source_api_response?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Name + * @description Name of the model. */ - use_cache?: boolean; + name?: string | null; /** - * @description SDXL Main model (UNet, VAE, CLIP1, CLIP2) to load - * @default null + * Path + * @description Path to the model. */ - model?: components["schemas"]["ModelIdentifierField"] | null; + path?: string | null; /** - * type - * @default sdxl_model_loader_input - * @constant + * Description + * @description Model description */ - type: "sdxl_model_loader_input"; - }; - /** - * Main Model - SDXL - * @description Loads an sdxl base model, outputting its submodels. - */ - SDXLModelLoaderInvocation: { + description?: string | null; + /** @description The base model. */ + base?: components["schemas"]["BaseModelType"] | null; + /** @description Type of model */ + type?: components["schemas"]["ModelType"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Key + * @description Database ID for this model */ - id: string; + key?: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Hash + * @description hash of model file */ - is_intermediate?: boolean; + hash?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * File Size + * @description Size of model file */ - use_cache?: boolean; + file_size?: number | null; /** - * @description SDXL Main model (UNet, VAE, CLIP1, CLIP2) to load - * @default null + * Format + * @description format of model file */ - model?: components["schemas"]["ModelIdentifierField"] | null; + format?: string | null; /** - * type - * @default sdxl_model_loader - * @constant + * Trigger Phrases + * @description Set of trigger phrases for this model */ - type: "sdxl_model_loader"; - }; - /** - * SDXLModelLoaderOutput - * @description SDXL base model loader output - */ - SDXLModelLoaderOutput: { + trigger_phrases?: string[] | null; /** - * UNet - * @description UNet (scheduler, LoRAs) + * Default Settings + * @description Default settings for this model */ - unet: components["schemas"]["UNetField"]; + default_settings?: components["schemas"]["MainModelDefaultSettings"] | components["schemas"]["LoraModelDefaultSettings"] | components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * CLIP 1 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Cpu Only + * @description Whether this model should run on CPU only */ - clip: components["schemas"]["CLIPField"]; + cpu_only?: boolean | null; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Variant + * @description The variant of the model. */ - clip2: components["schemas"]["CLIPField"]; + variant?: components["schemas"]["ModelVariantType"] | components["schemas"]["ClipVariantType"] | components["schemas"]["FluxVariantType"] | components["schemas"]["Flux2VariantType"] | components["schemas"]["ZImageVariantType"] | components["schemas"]["Qwen3VariantType"] | null; + /** @description The prediction type of the model. */ + prediction_type?: components["schemas"]["SchedulerPredictionType"] | null; /** - * VAE - * @description VAE + * Upcast Attention + * @description Whether to upcast attention. */ - vae: components["schemas"]["VAEField"]; + upcast_attention?: boolean | null; /** - * type - * @default sdxl_model_loader_output - * @constant + * Config Path + * @description Path to config file for model */ - type: "sdxl_model_loader_output"; + config_path?: string | null; }; /** - * SDXL Model To String - * @description Converts an SDXL Model to a JSONString + * ModelRecordOrderBy + * @description The order in which to return model summaries. + * @enum {string} */ - SDXLModelToStringInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + ModelRecordOrderBy: "default" | "type" | "base" | "name" | "format" | "size" | "created_at" | "updated_at" | "path"; + /** ModelRelationshipBatchRequest */ + ModelRelationshipBatchRequest: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Model Keys + * @description List of model keys to fetch related models for */ - use_cache?: boolean; + model_keys: string[]; + }; + /** ModelRelationshipCreateRequest */ + ModelRelationshipCreateRequest: { /** - * @description SDXL Main model (UNet, VAE, CLIP1, CLIP2) to load - * @default null + * Model Key 1 + * @description The key of the first model in the relationship */ - model?: components["schemas"]["ModelIdentifierField"] | null; + model_key_1: string; /** - * type - * @default sdxl_model_to_string - * @constant + * Model Key 2 + * @description The key of the second model in the relationship */ - type: "sdxl_model_to_string"; + model_key_2: string; }; /** - * Prompt - SDXL Refiner - * @description Parse prompt using compel package to conditioning. + * ModelRepoVariant + * @description Various hugging face variants on the diffusers format. + * @enum {string} */ - SDXLRefinerCompelPromptInvocation: { + ModelRepoVariant: "" | "fp16" | "fp32" | "onnx" | "openvino" | "flax"; + /** + * ModelSourceType + * @description Model source type. + * @enum {string} + */ + ModelSourceType: "path" | "url" | "hf_repo_id"; + /** + * ModelType + * @description Model type. + * @enum {string} + */ + ModelType: "onnx" | "main" | "vae" | "lora" | "control_lora" | "controlnet" | "embedding" | "ip_adapter" | "clip_vision" | "clip_embed" | "t2i_adapter" | "t5_encoder" | "qwen3_encoder" | "spandrel_image_to_image" | "siglip" | "flux_redux" | "llava_onevision" | "unknown"; + /** + * ModelVariantType + * @description Variant type. + * @enum {string} + */ + ModelVariantType: "normal" | "inpaint" | "depth"; + /** + * ModelsList + * @description Return list of configs. + */ + ModelsList: { + /** Models */ + models: (components["schemas"]["Main_Diffusers_SD1_Config"] | components["schemas"]["Main_Diffusers_SD2_Config"] | components["schemas"]["Main_Diffusers_SDXL_Config"] | components["schemas"]["Main_Diffusers_SDXLRefiner_Config"] | components["schemas"]["Main_Diffusers_SD3_Config"] | components["schemas"]["Main_Diffusers_FLUX_Config"] | components["schemas"]["Main_Diffusers_Flux2_Config"] | components["schemas"]["Main_Diffusers_CogView4_Config"] | components["schemas"]["Main_Diffusers_ZImage_Config"] | components["schemas"]["Main_Checkpoint_SD1_Config"] | components["schemas"]["Main_Checkpoint_SD2_Config"] | components["schemas"]["Main_Checkpoint_SDXL_Config"] | components["schemas"]["Main_Checkpoint_SDXLRefiner_Config"] | components["schemas"]["Main_Checkpoint_Flux2_Config"] | components["schemas"]["Main_Checkpoint_FLUX_Config"] | components["schemas"]["Main_Checkpoint_ZImage_Config"] | components["schemas"]["Main_BnBNF4_FLUX_Config"] | components["schemas"]["Main_GGUF_Flux2_Config"] | components["schemas"]["Main_GGUF_FLUX_Config"] | components["schemas"]["Main_GGUF_ZImage_Config"] | components["schemas"]["VAE_Checkpoint_SD1_Config"] | components["schemas"]["VAE_Checkpoint_SD2_Config"] | components["schemas"]["VAE_Checkpoint_SDXL_Config"] | components["schemas"]["VAE_Checkpoint_FLUX_Config"] | components["schemas"]["VAE_Checkpoint_Flux2_Config"] | components["schemas"]["VAE_Diffusers_SD1_Config"] | components["schemas"]["VAE_Diffusers_SDXL_Config"] | components["schemas"]["VAE_Diffusers_Flux2_Config"] | components["schemas"]["ControlNet_Checkpoint_SD1_Config"] | components["schemas"]["ControlNet_Checkpoint_SD2_Config"] | components["schemas"]["ControlNet_Checkpoint_SDXL_Config"] | components["schemas"]["ControlNet_Checkpoint_FLUX_Config"] | components["schemas"]["ControlNet_Checkpoint_ZImage_Config"] | components["schemas"]["ControlNet_Diffusers_SD1_Config"] | components["schemas"]["ControlNet_Diffusers_SD2_Config"] | components["schemas"]["ControlNet_Diffusers_SDXL_Config"] | components["schemas"]["ControlNet_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_SD1_Config"] | components["schemas"]["LoRA_LyCORIS_SD2_Config"] | components["schemas"]["LoRA_LyCORIS_SDXL_Config"] | components["schemas"]["LoRA_LyCORIS_Flux2_Config"] | components["schemas"]["LoRA_LyCORIS_FLUX_Config"] | components["schemas"]["LoRA_LyCORIS_ZImage_Config"] | components["schemas"]["LoRA_OMI_SDXL_Config"] | components["schemas"]["LoRA_OMI_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_SD1_Config"] | components["schemas"]["LoRA_Diffusers_SD2_Config"] | components["schemas"]["LoRA_Diffusers_SDXL_Config"] | components["schemas"]["LoRA_Diffusers_Flux2_Config"] | components["schemas"]["LoRA_Diffusers_FLUX_Config"] | components["schemas"]["LoRA_Diffusers_ZImage_Config"] | components["schemas"]["ControlLoRA_LyCORIS_FLUX_Config"] | components["schemas"]["T5Encoder_T5Encoder_Config"] | components["schemas"]["T5Encoder_BnBLLMint8_Config"] | components["schemas"]["Qwen3Encoder_Qwen3Encoder_Config"] | components["schemas"]["Qwen3Encoder_Checkpoint_Config"] | components["schemas"]["Qwen3Encoder_GGUF_Config"] | components["schemas"]["TI_File_SD1_Config"] | components["schemas"]["TI_File_SD2_Config"] | components["schemas"]["TI_File_SDXL_Config"] | components["schemas"]["TI_Folder_SD1_Config"] | components["schemas"]["TI_Folder_SD2_Config"] | components["schemas"]["TI_Folder_SDXL_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD1_Config"] | components["schemas"]["IPAdapter_InvokeAI_SD2_Config"] | components["schemas"]["IPAdapter_InvokeAI_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD1_Config"] | components["schemas"]["IPAdapter_Checkpoint_SD2_Config"] | components["schemas"]["IPAdapter_Checkpoint_SDXL_Config"] | components["schemas"]["IPAdapter_Checkpoint_FLUX_Config"] | components["schemas"]["T2IAdapter_Diffusers_SD1_Config"] | components["schemas"]["T2IAdapter_Diffusers_SDXL_Config"] | components["schemas"]["Spandrel_Checkpoint_Config"] | components["schemas"]["CLIPEmbed_Diffusers_G_Config"] | components["schemas"]["CLIPEmbed_Diffusers_L_Config"] | components["schemas"]["CLIPVision_Diffusers_Config"] | components["schemas"]["SigLIP_Diffusers_Config"] | components["schemas"]["FLUXRedux_Checkpoint_Config"] | components["schemas"]["LlavaOnevision_Diffusers_Config"] | components["schemas"]["Unknown_Config"])[]; + }; + /** + * Multiply Integers + * @description Multiplies two numbers + */ + MultiplyInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -36650,54 +21467,47 @@ export type components = { */ use_cache?: boolean; /** - * Style - * @description Prompt to be parsed by Compel to create a conditioning tensor - * @default - */ - style?: string; - /** - * Original Width - * @default 1024 - */ - original_width?: number; - /** - * Original Height - * @default 1024 + * A + * @description The first number + * @default 0 */ - original_height?: number; + a?: number; /** - * Crop Top + * B + * @description The second number * @default 0 */ - crop_top?: number; + b?: number; /** - * Crop Left - * @default 0 + * type + * @default mul + * @constant */ - crop_left?: number; + type: "mul"; + }; + /** NodeFieldValue */ + NodeFieldValue: { /** - * Aesthetic Score - * @description The aesthetic score to apply to the conditioning tensor - * @default 6 + * Node Path + * @description The node into which this batch data item will be substituted. */ - aesthetic_score?: number; + node_path: string; /** - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null + * Field Name + * @description The field into which this batch data item will be substituted. */ - clip2?: components["schemas"]["CLIPField"] | null; + field_name: string; /** - * type - * @default sdxl_refiner_compel_prompt - * @constant + * Value + * @description The value to substitute into the node/field. */ - type: "sdxl_refiner_compel_prompt"; + value: string | number | components["schemas"]["ImageField"]; }; /** - * Refiner Model - SDXL - * @description Loads an sdxl refiner model, outputting its submodels. + * Create Latent Noise + * @description Generates latent noise. */ - SDXLRefinerModelLoaderInvocation: { + NoiseInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -36716,54 +21526,65 @@ export type components = { */ use_cache?: boolean; /** - * @description SDXL Refiner Main Modde (UNet, VAE, CLIP2) to load - * @default null + * Seed + * @description Seed for random number generation + * @default 0 */ - model?: components["schemas"]["ModelIdentifierField"] | null; + seed?: number; + /** + * Width + * @description Width of output (px) + * @default 512 + */ + width?: number; + /** + * Height + * @description Height of output (px) + * @default 512 + */ + height?: number; + /** + * Use Cpu + * @description Use CPU for noise generation (for reproducible results across platforms) + * @default true + */ + use_cpu?: boolean; /** * type - * @default sdxl_refiner_model_loader + * @default noise * @constant */ - type: "sdxl_refiner_model_loader"; + type: "noise"; }; /** - * SDXLRefinerModelLoaderOutput - * @description SDXL refiner model loader output + * NoiseOutput + * @description Invocation noise output */ - SDXLRefinerModelLoaderOutput: { - /** - * UNet - * @description UNet (scheduler, LoRAs) - */ - unet: components["schemas"]["UNetField"]; + NoiseOutput: { + /** @description Noise tensor */ + noise: components["schemas"]["LatentsField"]; /** - * CLIP 2 - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Width + * @description Width of output (px) */ - clip2: components["schemas"]["CLIPField"]; + width: number; /** - * VAE - * @description VAE + * Height + * @description Height of output (px) */ - vae: components["schemas"]["VAEField"]; + height: number; /** * type - * @default sdxl_refiner_model_loader_output + * @default noise_output * @constant */ - type: "sdxl_refiner_model_loader_output"; + type: "noise_output"; }; /** - * SQLiteDirection - * @enum {string} - */ - SQLiteDirection: "ASC" | "DESC"; - /** - * Save Image - * @description Saves an image. Unlike an image primitive, this invocation stores a copy of the image. + * Normal Map + * @description Generates a normal map. */ - SaveImageInvocation: { + NormalMapInvocation: { /** * @description The board to save the image to * @default null @@ -36788,7 +21609,7 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default false + * @default true */ use_cache?: boolean; /** @@ -36798,162 +21619,118 @@ export type components = { image?: components["schemas"]["ImageField"] | null; /** * type - * @default save_image + * @default normal_map * @constant */ - type: "save_image"; + type: "normal_map"; }; - /** - * Scale Latents - * @description Scales latents by a given factor. - */ - ScaleLatentsInvocation: { + /** OffsetPaginatedResults[BoardDTO] */ + OffsetPaginatedResults_BoardDTO_: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Limit + * @description Limit of items to get */ - id: string; + limit: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Offset + * @description Offset from which to retrieve items */ - is_intermediate?: boolean; + offset: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Total + * @description Total number of items in result */ - use_cache?: boolean; + total: number; /** - * @description Latents tensor - * @default null + * Items + * @description Items */ - latents?: components["schemas"]["LatentsField"] | null; + items: components["schemas"]["BoardDTO"][]; + }; + /** OffsetPaginatedResults[ImageDTO] */ + OffsetPaginatedResults_ImageDTO_: { /** - * Scale Factor - * @description The factor by which to scale - * @default null + * Limit + * @description Limit of items to get */ - scale_factor?: number | null; + limit: number; /** - * Mode - * @description Interpolation mode - * @default bilinear - * @enum {string} + * Offset + * @description Offset from which to retrieve items */ - mode?: "nearest" | "linear" | "bilinear" | "bicubic" | "trilinear" | "area" | "nearest-exact"; + offset: number; /** - * Antialias - * @description Whether or not to apply antialiasing (bilinear or bicubic only) - * @default false + * Total + * @description Total number of items in result */ - antialias?: boolean; + total: number; /** - * type - * @default lscale - * @constant + * Items + * @description Items */ - type: "lscale"; + items: components["schemas"]["ImageDTO"][]; }; /** - * Scheduler - * @description Selects a scheduler. + * OrphanedModelInfo + * @description Information about an orphaned model directory. */ - SchedulerInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; + OrphanedModelInfo: { /** - * Scheduler - * @description Scheduler to use during inference - * @default euler - * @enum {string} + * Path + * @description Relative path to the orphaned directory from models root */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + path: string; /** - * type - * @default scheduler - * @constant + * Absolute Path + * @description Absolute path to the orphaned directory */ - type: "scheduler"; - }; - /** SchedulerOutput */ - SchedulerOutput: { + absolute_path: string; /** - * Scheduler - * @description Scheduler to use during inference - * @enum {string} + * Files + * @description List of model files in this directory */ - scheduler: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + files: string[]; /** - * type - * @default scheduler_output - * @constant + * Size Bytes + * @description Total size of all files in bytes */ - type: "scheduler_output"; + size_bytes: number; }; /** - * SchedulerPredictionType - * @description Scheduler prediction type. - * @enum {string} - */ - SchedulerPredictionType: "epsilon" | "v_prediction" | "sample"; - /** - * Scheduler To String - * @description Converts a Scheduler to a string + * OutputFieldJSONSchemaExtra + * @description Extra attributes to be added to input fields and their OpenAPI schema. Used by the workflow editor + * during schema parsing and UI rendering. */ - SchedulerToStringInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + OutputFieldJSONSchemaExtra: { + field_kind: components["schemas"]["FieldKind"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. + * Ui Hidden * @default false */ - is_intermediate?: boolean; + ui_hidden: boolean; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Ui Order + * @default null */ - use_cache?: boolean; + ui_order: number | null; + /** @default null */ + ui_type: components["schemas"]["UIType"] | null; + }; + /** + * PBR Maps + * @description Generate Normal, Displacement and Roughness Map from a given image + */ + PBRMapsInvocation: { /** - * Scheduler - * @description Scheduler to use during inference - * @default euler - * @enum {string} + * @description The board to save the image to + * @default null */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + board?: components["schemas"]["BoardField"] | null; /** - * type - * @default scheduler_to_string - * @constant + * @description Optional metadata to be saved with the image + * @default null */ - type: "scheduler_to_string"; - }; - /** - * Scheduler Toggle - * @description Allows boolean selection between two separate scheduler inputs - */ - SchedulerToggleInvocation: { + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -36972,127 +21749,87 @@ export type components = { */ use_cache?: boolean; /** - * Use Second - * @description Use 2nd Input - * @default false + * @description Input image + * @default null */ - use_second?: boolean; + image?: components["schemas"]["ImageField"] | null; /** - * Scheduler1 - * @description First Scheduler Input - * @default null + * Tile Size + * @description Tile size + * @default 512 */ - scheduler1?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; + tile_size?: number; /** - * Scheduler2 - * @description Second Scheduler Input - * @default null + * Border Mode + * @description Border mode to apply to eliminate any artifacts or seams + * @default none + * @enum {string} */ - scheduler2?: ("ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd") | null; + border_mode?: "none" | "seamless" | "mirror" | "replicate"; /** * type - * @default scheduler_toggle + * @default pbr_maps * @constant */ - type: "scheduler_toggle"; + type: "pbr_maps"; }; - /** - * Main Model - SD3 - * @description Loads a SD3 base model, outputting its submodels. - */ - Sd3ModelLoaderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; - /** - * Use Cache - * @description Whether or not to use the cache - * @default true - */ - use_cache?: boolean; - /** @description SD3 model (MMDiTX) to load */ - model: components["schemas"]["ModelIdentifierField"]; - /** - * T5 Encoder - * @description T5 tokenizer and text encoder - * @default null - */ - t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; + /** PBRMapsOutput */ + PBRMapsOutput: { /** - * CLIP L Encoder - * @description CLIP Embed loader + * @description The generated normal map * @default null */ - clip_l_model?: components["schemas"]["ModelIdentifierField"] | null; + normal_map: components["schemas"]["ImageField"]; /** - * CLIP G Encoder - * @description CLIP-G Embed loader + * @description The generated roughness map * @default null */ - clip_g_model?: components["schemas"]["ModelIdentifierField"] | null; + roughness_map: components["schemas"]["ImageField"]; /** - * VAE - * @description VAE model to load + * @description The generated displacement map * @default null */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + displacement_map: components["schemas"]["ImageField"]; /** * type - * @default sd3_model_loader + * @default pbr_maps-output * @constant */ - type: "sd3_model_loader"; - }; - /** - * Sd3ModelLoaderOutput - * @description SD3 base model loader output. - */ - Sd3ModelLoaderOutput: { - /** - * Transformer - * @description Transformer - */ - transformer: components["schemas"]["TransformerField"]; + type: "pbr_maps-output"; + }; + /** PaginatedResults[WorkflowRecordListItemWithThumbnailDTO] */ + PaginatedResults_WorkflowRecordListItemWithThumbnailDTO_: { /** - * CLIP L - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Page + * @description Current Page */ - clip_l: components["schemas"]["CLIPField"]; + page: number; /** - * CLIP G - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * Pages + * @description Total number of pages */ - clip_g: components["schemas"]["CLIPField"]; + pages: number; /** - * T5 Encoder - * @description T5 tokenizer and text encoder + * Per Page + * @description Number of items per page */ - t5_encoder: components["schemas"]["T5EncoderField"]; + per_page: number; /** - * VAE - * @description VAE + * Total + * @description Total number of items in result */ - vae: components["schemas"]["VAEField"]; + total: number; /** - * type - * @default sd3_model_loader_output - * @constant + * Items + * @description Items */ - type: "sd3_model_loader_output"; + items: components["schemas"]["WorkflowRecordListItemWithThumbnailDTO"][]; }; /** - * SD3 Model To String - * @description Converts an SD3 Model to a JSONString + * Pair Tile with Image + * @description Pair an image with its tile properties. */ - Sd3ModelToStringInvocation: { + PairTileImageInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -37111,22 +21848,50 @@ export type components = { */ use_cache?: boolean; /** - * @description SD3 model (MMDiTX) to load + * @description The tile image. * @default null */ - model?: components["schemas"]["ModelIdentifierField"] | null; + image?: components["schemas"]["ImageField"] | null; + /** + * @description The tile properties. + * @default null + */ + tile?: components["schemas"]["Tile"] | null; /** * type - * @default sd3_model_to_string + * @default pair_tile_image + * @constant + */ + type: "pair_tile_image"; + }; + /** PairTileImageOutput */ + PairTileImageOutput: { + /** @description A tile description with its corresponding image. */ + tile_with_image: components["schemas"]["TileWithImage"]; + /** + * type + * @default pair_tile_image_output * @constant */ - type: "sd3_model_to_string"; + type: "pair_tile_image_output"; }; /** - * Prompt - SD3 - * @description Encodes and preps a prompt for a SD3 image. + * Paste Image into Bounding Box + * @description Paste the source image into the target image at the given bounding box. + * + * The source image must be the same size as the bounding box, and the bounding box must fit within the target image. */ - Sd3TextEncoderInvocation: { + PasteImageIntoBoundingBoxInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -37145,41 +21910,42 @@ export type components = { */ use_cache?: boolean; /** - * CLIP L - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count - * @default null - */ - clip_l?: components["schemas"]["CLIPField"] | null; - /** - * CLIP G - * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @description The image to paste * @default null */ - clip_g?: components["schemas"]["CLIPField"] | null; + source_image?: components["schemas"]["ImageField"] | null; /** - * T5Encoder - * @description T5 tokenizer and text encoder + * @description The image to paste into * @default null */ - t5_encoder?: components["schemas"]["T5EncoderField"] | null; + target_image?: components["schemas"]["ImageField"] | null; /** - * Prompt - * @description Text prompt to encode. + * @description The bounding box to paste the image into * @default null */ - prompt?: string | null; + bounding_box?: components["schemas"]["BoundingBoxField"] | null; /** * type - * @default sd3_text_encoder + * @default paste_image_into_bounding_box * @constant */ - type: "sd3_text_encoder"; + type: "paste_image_into_bounding_box"; }; /** - * Apply Seamless - SD1.5, SDXL - * @description Applies the seamless transformation to the Model UNet and VAE. + * PiDiNet Edge Detection + * @description Generates an edge map using PiDiNet. */ - SeamlessModeInvocation: { + PiDiNetEdgeDetectionInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -37198,65 +21964,76 @@ export type components = { */ use_cache?: boolean; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null - */ - unet?: components["schemas"]["UNetField"] | null; - /** - * VAE - * @description VAE model to load + * @description The image to process * @default null */ - vae?: components["schemas"]["VAEField"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Seamless Y - * @description Specify whether Y axis is seamless - * @default true + * Quantize Edges + * @description Whether or not to use safe mode + * @default false */ - seamless_y?: boolean; + quantize_edges?: boolean; /** - * Seamless X - * @description Specify whether X axis is seamless - * @default true + * Scribble + * @description Whether or not to use scribble mode + * @default false */ - seamless_x?: boolean; + scribble?: boolean; /** * type - * @default seamless + * @default pidi_edge_detection * @constant */ - type: "seamless"; + type: "pidi_edge_detection"; + }; + /** PresetData */ + PresetData: { + /** + * Positive Prompt + * @description Positive prompt + */ + positive_prompt: string; + /** + * Negative Prompt + * @description Negative prompt + */ + negative_prompt: string; }; /** - * SeamlessModeOutput - * @description Modified Seamless Model output + * PresetType + * @enum {string} */ - SeamlessModeOutput: { + PresetType: "user" | "default"; + /** + * ProgressImage + * @description The progress image sent intermittently during processing + */ + ProgressImage: { /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Width + * @description The effective width of the image in pixels */ - unet: components["schemas"]["UNetField"] | null; + width: number; /** - * VAE - * @description VAE - * @default null + * Height + * @description The effective height of the image in pixels */ - vae: components["schemas"]["VAEField"] | null; + height: number; /** - * type - * @default seamless_output - * @constant + * Dataurl + * @description The image data as a b64 data URL */ - type: "seamless_output"; + dataURL: string; }; /** - * Segment Anything - * @description Runs a Segment Anything Model (SAM or SAM2). + * Prompt Template + * @description Applies a Style Preset template to positive and negative prompts. + * + * Select a Style Preset and provide positive/negative prompts. The node replaces + * {prompt} placeholders in the template with your input prompts. */ - SegmentAnythingInvocation: { + PromptTemplateInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -37275,54 +22052,56 @@ export type components = { */ use_cache?: boolean; /** - * Model - * @description The Segment Anything model to use (SAM or SAM2). + * @description The Style Preset to use as a template * @default null */ - model?: ("segment-anything-base" | "segment-anything-large" | "segment-anything-huge" | "segment-anything-2-tiny" | "segment-anything-2-small" | "segment-anything-2-base" | "segment-anything-2-large") | null; + style_preset?: components["schemas"]["StylePresetField"] | null; /** - * @description The image to segment. - * @default null + * Positive Prompt + * @description The positive prompt to insert into the template's {prompt} placeholder + * @default */ - image?: components["schemas"]["ImageField"] | null; + positive_prompt?: string; /** - * Bounding Boxes - * @description The bounding boxes to prompt the model with. - * @default null + * Negative Prompt + * @description The negative prompt to insert into the template's {prompt} placeholder + * @default */ - bounding_boxes?: components["schemas"]["BoundingBoxField"][] | null; + negative_prompt?: string; /** - * Point Lists - * @description The list of point lists to prompt the model with. Each list of points represents a single object. - * @default null + * type + * @default prompt_template + * @constant */ - point_lists?: components["schemas"]["SAMPointsField"][] | null; + type: "prompt_template"; + }; + /** + * PromptTemplateOutput + * @description Output for the Prompt Template node + */ + PromptTemplateOutput: { /** - * Apply Polygon Refinement - * @description Whether to apply polygon refinement to the masks. This will smooth the edges of the masks slightly and ensure that each mask consists of a single closed polygon (before merging). - * @default true + * Positive Prompt + * @description The positive prompt with the template applied */ - apply_polygon_refinement?: boolean; + positive_prompt: string; /** - * Mask Filter - * @description The filtering to apply to the detected masks before merging them into a final output. - * @default all - * @enum {string} + * Negative Prompt + * @description The negative prompt with the template applied */ - mask_filter?: "all" | "largest" | "highest_box_score"; + negative_prompt: string; /** * type - * @default segment_anything + * @default prompt_template_output * @constant */ - type: "segment_anything"; + type: "prompt_template_output"; }; /** - * Separate Prompt and Seed Vector - * @description Parses a JSON string representing a list of two strings, - * outputting each string separately. + * Prompts from File + * @description Loads prompts from a text file */ - SeparatePromptAndSeedVectorInvocation: { + PromptsFromFileInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -37341,466 +22120,378 @@ export type components = { */ use_cache?: boolean; /** - * Pair Input - * @description JSON string of a list containing exactly two strings, e.g., '["string one", "string two"]' - * @default ["", ""] - */ - pair_input?: string; - /** - * type - * @default separate_prompt_and_seed_vector - * @constant - */ - type: "separate_prompt_and_seed_vector"; - }; - /** SessionProcessorStatus */ - SessionProcessorStatus: { - /** - * Is Started - * @description Whether the session processor is started - */ - is_started: boolean; - /** - * Is Processing - * @description Whether a session is being processed + * File Path + * @description Path to prompt text file + * @default null */ - is_processing: boolean; - }; - /** - * SessionQueueAndProcessorStatus - * @description The overall status of session queue and processor - */ - SessionQueueAndProcessorStatus: { - queue: components["schemas"]["SessionQueueStatus"]; - processor: components["schemas"]["SessionProcessorStatus"]; - }; - /** SessionQueueCountsByDestination */ - SessionQueueCountsByDestination: { + file_path?: string | null; /** - * Queue Id - * @description The ID of the queue + * Pre Prompt + * @description String to prepend to each prompt + * @default null */ - queue_id: string; + pre_prompt?: string | null; /** - * Destination - * @description The destination of queue items included in this status + * Post Prompt + * @description String to append to each prompt + * @default null */ - destination: string; + post_prompt?: string | null; /** - * Pending - * @description Number of queue items with status 'pending' for the destination + * Start Line + * @description Line in the file to start start from + * @default 1 */ - pending: number; + start_line?: number; /** - * In Progress - * @description Number of queue items with status 'in_progress' for the destination + * Max Prompts + * @description Max lines to read from file (0=all) + * @default 1 */ - in_progress: number; + max_prompts?: number; /** - * Completed - * @description Number of queue items with status 'complete' for the destination + * type + * @default prompt_from_file + * @constant */ - completed: number; + type: "prompt_from_file"; + }; + /** + * PruneResult + * @description Result of pruning the session queue + */ + PruneResult: { /** - * Failed - * @description Number of queue items with status 'error' for the destination + * Deleted + * @description Number of queue items deleted */ - failed: number; + deleted: number; + }; + /** + * QueueClearedEvent + * @description Event model for queue_cleared + */ + QueueClearedEvent: { /** - * Canceled - * @description Number of queue items with status 'canceled' for the destination + * Timestamp + * @description The timestamp of the event */ - canceled: number; + timestamp: number; /** - * Total - * @description Total number of queue items for the destination + * Queue Id + * @description The ID of the queue */ - total: number; + queue_id: string; }; /** - * SessionQueueItem - * @description Session queue item without the full graph. Used for serialization. + * QueueItemStatusChangedEvent + * @description Event model for queue_item_status_changed */ - SessionQueueItem: { + QueueItemStatusChangedEvent: { /** - * Item Id - * @description The identifier of the session queue item + * Timestamp + * @description The timestamp of the event */ - item_id: number; + timestamp: number; /** - * Status - * @description The status of this queue item - * @default pending - * @enum {string} + * Queue Id + * @description The ID of the queue */ - status: "pending" | "in_progress" | "completed" | "failed" | "canceled"; + queue_id: string; /** - * Priority - * @description The priority of this queue item - * @default 0 + * Item Id + * @description The ID of the queue item */ - priority: number; + item_id: number; /** * Batch Id - * @description The ID of the batch associated with this queue item + * @description The ID of the queue batch */ batch_id: string; /** * Origin - * @description The origin of this queue item. This data is used by the frontend to determine how to handle results. + * @description The origin of the queue item + * @default null */ - origin?: string | null; + origin: string | null; /** * Destination - * @description The origin of this queue item. This data is used by the frontend to determine how to handle results - */ - destination?: string | null; - /** - * Session Id - * @description The ID of the session associated with this queue item. The session doesn't exist in graph_executions until the queue item is executed. - */ - session_id: string; - /** - * Error Type - * @description The error type if this queue item errored - */ - error_type?: string | null; - /** - * Error Message - * @description The error message if this queue item errored - */ - error_message?: string | null; - /** - * Error Traceback - * @description The error traceback if this queue item errored - */ - error_traceback?: string | null; - /** - * Created At - * @description When this queue item was created - */ - created_at: string; - /** - * Updated At - * @description When this queue item was updated - */ - updated_at: string; - /** - * Started At - * @description When this queue item was started - */ - started_at?: string | null; - /** - * Completed At - * @description When this queue item was completed - */ - completed_at?: string | null; - /** - * Queue Id - * @description The id of the queue with which this item is associated + * @description The destination of the queue item + * @default null */ - queue_id: string; + destination: string | null; /** * User Id - * @description The id of the user who created this queue item + * @description The ID of the user who created the queue item * @default system */ - user_id?: string; - /** - * User Display Name - * @description The display name of the user who created this queue item, if available - */ - user_display_name?: string | null; - /** - * User Email - * @description The email of the user who created this queue item, if available - */ - user_email?: string | null; - /** - * Field Values - * @description The field values that were used for this queue item - */ - field_values?: components["schemas"]["NodeFieldValue"][] | null; - /** - * Retried From Item Id - * @description The item_id of the queue item that this item was retried from - */ - retried_from_item_id?: number | null; - /** @description The fully-populated session to be executed */ - session: components["schemas"]["GraphExecutionState"]; - /** @description The workflow associated with this queue item */ - workflow?: components["schemas"]["WorkflowWithoutID"] | null; - }; - /** SessionQueueStatus */ - SessionQueueStatus: { - /** - * Queue Id - * @description The ID of the queue - */ - queue_id: string; - /** - * Item Id - * @description The current queue item id - */ - item_id: number | null; - /** - * Batch Id - * @description The current queue item's batch id - */ - batch_id: string | null; + user_id: string; /** - * Session Id - * @description The current queue item's session id + * Status + * @description The new status of the queue item + * @enum {string} */ - session_id: string | null; + status: "pending" | "in_progress" | "completed" | "failed" | "canceled"; /** - * Pending - * @description Number of queue items with status 'pending' + * Error Type + * @description The error type, if any + * @default null */ - pending: number; + error_type: string | null; /** - * In Progress - * @description Number of queue items with status 'in_progress' + * Error Message + * @description The error message, if any + * @default null */ - in_progress: number; - /** - * Completed - * @description Number of queue items with status 'complete' + error_message: string | null; + /** + * Error Traceback + * @description The error traceback, if any + * @default null */ - completed: number; + error_traceback: string | null; /** - * Failed - * @description Number of queue items with status 'error' + * Created At + * @description The timestamp when the queue item was created */ - failed: number; + created_at: string; /** - * Canceled - * @description Number of queue items with status 'canceled' + * Updated At + * @description The timestamp when the queue item was last updated */ - canceled: number; + updated_at: string; /** - * Total - * @description Total number of queue items + * Started At + * @description The timestamp when the queue item was started + * @default null */ - total: number; + started_at: string | null; /** - * User Pending - * @description Number of queue items with status 'pending' for the current user + * Completed At + * @description The timestamp when the queue item was completed + * @default null */ - user_pending?: number | null; + completed_at: string | null; + /** @description The status of the batch */ + batch_status: components["schemas"]["BatchStatus"]; + /** @description The status of the queue */ + queue_status: components["schemas"]["SessionQueueStatus"]; /** - * User In Progress - * @description Number of queue items with status 'in_progress' for the current user + * Session Id + * @description The ID of the session (aka graph execution state) */ - user_in_progress?: number | null; + session_id: string; }; /** - * SetupRequest - * @description Request body for initial admin setup. + * QueueItemsRetriedEvent + * @description Event model for queue_items_retried */ - SetupRequest: { + QueueItemsRetriedEvent: { /** - * Email - * @description Admin email address + * Timestamp + * @description The timestamp of the event */ - email: string; + timestamp: number; /** - * Display Name - * @description Admin display name + * Queue Id + * @description The ID of the queue */ - display_name?: string | null; + queue_id: string; /** - * Password - * @description Admin password + * Retried Item Ids + * @description The IDs of the queue items that were retried */ - password: string; + retried_item_ids: number[]; }; /** - * SetupResponse - * @description Response from successful admin setup. + * Qwen3EncoderField + * @description Field for Qwen3 text encoder used by Z-Image models. */ - SetupResponse: { + Qwen3EncoderField: { + /** @description Info to load tokenizer submodel */ + tokenizer: components["schemas"]["ModelIdentifierField"]; + /** @description Info to load text_encoder submodel */ + text_encoder: components["schemas"]["ModelIdentifierField"]; /** - * Success - * @description Whether setup was successful + * Loras + * @description LoRAs to apply on model loading */ - success: boolean; - /** @description Created admin user information */ - user: components["schemas"]["UserDTO"]; + loras?: components["schemas"]["LoRAField"][]; }; /** - * SetupStatusResponse - * @description Response for setup status check. + * Qwen3Encoder_Checkpoint_Config + * @description Configuration for single-file Qwen3 Encoder models (safetensors). */ - SetupStatusResponse: { + Qwen3Encoder_Checkpoint_Config: { /** - * Setup Required - * @description Whether initial setup is required + * Key + * @description A unique key for this model. */ - setup_required: boolean; + key: string; /** - * Multiuser Enabled - * @description Whether multiuser mode is enabled + * Hash + * @description The hash of the model file(s). */ - multiuser_enabled: boolean; + hash: string; /** - * Strict Password Checking - * @description Whether strict password requirements are enforced + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - strict_password_checking: boolean; - }; - /** - * Shadows/Highlights/Midtones - * @description Extract a Shadows/Highlights/Midtones mask from an image. - * - * Extract three masks (with adjustable hard or soft thresholds) representing shadows, midtones, and highlights regions of an image. - */ - ShadowsHighlightsMidtonesMaskInvocation: { + path: string; /** - * @description The board to save the image to - * @default null + * File Size + * @description The size of the model in bytes. */ - board?: components["schemas"]["BoardField"] | null; + file_size: number; /** - * @description Optional metadata to be saved with the image - * @default null + * Name + * @description Name of the model. */ - metadata?: components["schemas"]["MetadataField"] | null; + name: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Description + * @description Model description */ - id: string; + description: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - is_intermediate?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - use_cache?: boolean; + source_api_response: string | null; /** - * @description Image from which to extract mask - * @default null + * Cover Image + * @description Url for image to preview model */ - image?: components["schemas"]["ImageField"] | null; + cover_image: string | null; /** - * Invert Output - * @description Off: white on black / On: black on white - * @default true + * Config Path + * @description Path to the config for this model, if any. */ - invert_output?: boolean; + config_path: string | null; /** - * Highlight Threshold - * @description Threshold beyond which mask values will be at extremum - * @default 0.75 + * Base + * @default any + * @constant */ - highlight_threshold?: number; + base: "any"; /** - * Upper Mid Threshold - * @description Threshold to which to extend mask border by 0..1 gradient - * @default 0.7 + * Type + * @default qwen3_encoder + * @constant */ - upper_mid_threshold?: number; + type: "qwen3_encoder"; /** - * Lower Mid Threshold - * @description Threshold to which to extend mask border by 0..1 gradient - * @default 0.3 + * Format + * @default checkpoint + * @constant */ - lower_mid_threshold?: number; + format: "checkpoint"; /** - * Shadow Threshold - * @description Threshold beyond which mask values will be at extremum - * @default 0.25 + * Cpu Only + * @description Whether this model should run on CPU only */ - shadow_threshold?: number; + cpu_only: boolean | null; + /** @description Qwen3 model size variant (4B or 8B) */ + variant: components["schemas"]["Qwen3VariantType"]; + }; + /** + * Qwen3Encoder_GGUF_Config + * @description Configuration for GGUF-quantized Qwen3 Encoder models. + */ + Qwen3Encoder_GGUF_Config: { /** - * Mask Expand Or Contract - * @description Pixels to grow (or shrink) the mask areas - * @default 0 + * Key + * @description A unique key for this model. */ - mask_expand_or_contract?: number; + key: string; /** - * Mask Blur - * @description Gaussian blur radius to apply to the masks - * @default 0 + * Hash + * @description The hash of the model file(s). */ - mask_blur?: number; + hash: string; /** - * type - * @default shmmask - * @constant + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - type: "shmmask"; - }; - /** ShadowsHighlightsMidtonesMasksOutput */ - ShadowsHighlightsMidtonesMasksOutput: { - /** @description Soft-edged highlights mask */ - highlights_mask: components["schemas"]["ImageField"]; - /** @description Soft-edged midtones mask */ - midtones_mask: components["schemas"]["ImageField"]; - /** @description Soft-edged shadows mask */ - shadows_mask: components["schemas"]["ImageField"]; + path: string; /** - * Width - * @description Width of the input/outputs + * File Size + * @description The size of the model in bytes. */ - width: number; + file_size: number; /** - * Height - * @description Height of the input/outputs + * Name + * @description Name of the model. */ - height: number; + name: string; /** - * type - * @default shmmask_output - * @constant + * Description + * @description Model description */ - type: "shmmask_output"; - }; - /** - * Show Image - * @description Displays a provided image using the OS image viewer, and passes it forward in the pipeline. - */ - ShowImageInvocation: { + description: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source + * @description The original source of the model (path, URL or repo_id). */ - id: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - is_intermediate?: boolean; + source_api_response: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Cover Image + * @description Url for image to preview model */ - use_cache?: boolean; + cover_image: string | null; /** - * @description The image to show - * @default null + * Config Path + * @description Path to the config for this model, if any. + */ + config_path: string | null; + /** + * Base + * @default any + * @constant + */ + base: "any"; + /** + * Type + * @default qwen3_encoder + * @constant */ - image?: components["schemas"]["ImageField"] | null; + type: "qwen3_encoder"; /** - * type - * @default show_image + * Format + * @default gguf_quantized * @constant */ - type: "show_image"; + format: "gguf_quantized"; + /** + * Cpu Only + * @description Whether this model should run on CPU only + */ + cpu_only: boolean | null; + /** @description Qwen3 model size variant (4B or 8B) */ + variant: components["schemas"]["Qwen3VariantType"]; }; /** - * SigLIP_Diffusers_Config - * @description Model config for SigLIP. + * Qwen3Encoder_Qwen3Encoder_Config + * @description Configuration for Qwen3 Encoder models in a diffusers-like format. + * + * The model weights are expected to be in a folder called text_encoder inside the model directory, + * compatible with Qwen2VLForConditionalGeneration or similar architectures used by Z-Image. */ - SigLIP_Diffusers_Config: { + Qwen3Encoder_Qwen3Encoder_Config: { /** * Key * @description A unique key for this model. @@ -37849,46 +22540,42 @@ export type components = { */ cover_image: string | null; /** - * Format - * @default diffusers + * Base + * @default any * @constant */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + base: "any"; /** * Type - * @default siglip + * @default qwen3_encoder * @constant */ - type: "siglip"; + type: "qwen3_encoder"; /** - * Base - * @default any + * Format + * @default qwen3_encoder * @constant */ - base: "any"; + format: "qwen3_encoder"; /** * Cpu Only * @description Whether this model should run on CPU only */ cpu_only: boolean | null; + /** @description Qwen3 model size variant (4B or 8B) */ + variant: components["schemas"]["Qwen3VariantType"]; }; /** - * Image-to-Image (Autoscale) - * @description Run any spandrel image-to-image model (https://github.com/chaiNNer-org/spandrel) until the target scale is reached. + * Qwen3VariantType + * @description Qwen3 text encoder variants based on model size. + * @enum {string} */ - SpandrelImageToImageAutoscaleInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + Qwen3VariantType: "qwen3_4b" | "qwen3_8b"; + /** + * Random Float + * @description Outputs a single random float + */ + RandomFloatInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -37903,60 +22590,80 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * @description The input image - * @default null + * Low + * @description The inclusive low value + * @default 0 */ - image?: components["schemas"]["ImageField"] | null; + low?: number; /** - * Image-to-Image Model - * @description Image-to-Image model - * @default null + * High + * @description The exclusive high value + * @default 1 */ - image_to_image_model?: components["schemas"]["ModelIdentifierField"] | null; + high?: number; /** - * Tile Size - * @description The tile size for tiled image-to-image. Set to 0 to disable tiling. - * @default 512 + * Decimals + * @description The number of decimal places to round to + * @default 2 */ - tile_size?: number; + decimals?: number; /** * type - * @default spandrel_image_to_image_autoscale + * @default rand_float * @constant */ - type: "spandrel_image_to_image_autoscale"; + type: "rand_float"; + }; + /** + * Random Integer + * @description Outputs a single random integer. + */ + RandomIntInvocation: { /** - * Scale - * @description The final scale of the output image. If the model does not upscale the image, this will be ignored. - * @default 4 + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - scale?: number; + id: string; /** - * Fit To Multiple Of 8 - * @description If true, the output image will be resized to the nearest multiple of 8 in both dimensions. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. * @default false */ - fit_to_multiple_of_8?: boolean; - }; - /** - * Image-to-Image - * @description Run any spandrel image-to-image model (https://github.com/chaiNNer-org/spandrel). - */ - SpandrelImageToImageInvocation: { + is_intermediate?: boolean; /** - * @description The board to save the image to - * @default null + * Use Cache + * @description Whether or not to use the cache + * @default false */ - board?: components["schemas"]["BoardField"] | null; + use_cache?: boolean; /** - * @description Optional metadata to be saved with the image - * @default null + * Low + * @description The inclusive low value + * @default 0 */ - metadata?: components["schemas"]["MetadataField"] | null; + low?: number; + /** + * High + * @description The exclusive high value + * @default 2147483647 + */ + high?: number; + /** + * type + * @default rand_int + * @constant + */ + type: "rand_int"; + }; + /** + * Random Range + * @description Creates a collection of random numbers + */ + RandomRangeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -37971,310 +22678,298 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * @description The input image - * @default null + * Low + * @description The inclusive low value + * @default 0 */ - image?: components["schemas"]["ImageField"] | null; + low?: number; /** - * Image-to-Image Model - * @description Image-to-Image model - * @default null + * High + * @description The exclusive high value + * @default 2147483647 */ - image_to_image_model?: components["schemas"]["ModelIdentifierField"] | null; + high?: number; /** - * Tile Size - * @description The tile size for tiled image-to-image. Set to 0 to disable tiling. - * @default 512 + * Size + * @description The number of values to generate + * @default 1 */ - tile_size?: number; + size?: number; + /** + * Seed + * @description The seed for the RNG (omit for random) + * @default 0 + */ + seed?: number; /** * type - * @default spandrel_image_to_image + * @default random_range * @constant */ - type: "spandrel_image_to_image"; + type: "random_range"; }; /** - * Spandrel_Checkpoint_Config - * @description Model config for Spandrel Image to Image models. + * Integer Range + * @description Creates a range of numbers from start to stop with step */ - Spandrel_Checkpoint_Config: { + RangeInvocation: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Start + * @description The start of the range + * @default 0 */ - file_size: number; + start?: number; /** - * Name - * @description Name of the model. + * Stop + * @description The stop of the range + * @default 10 */ - name: string; + stop?: number; /** - * Description - * @description Model description + * Step + * @description The step of the range + * @default 1 */ - description: string | null; + step?: number; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default range + * @constant + */ + type: "range"; + }; + /** + * Integer Range of Size + * @description Creates a range from start to start + (size * step) incremented by step + */ + RangeOfSizeInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Start + * @description The start of the range + * @default 0 */ - cover_image: string | null; + start?: number; /** - * Base - * @default any - * @constant + * Size + * @description The number of values + * @default 1 */ - base: "any"; + size?: number; /** - * Type - * @default spandrel_image_to_image - * @constant + * Step + * @description The step of the range + * @default 1 */ - type: "spandrel_image_to_image"; + step?: number; /** - * Format - * @default checkpoint + * type + * @default range_of_size * @constant */ - format: "checkpoint"; + type: "range_of_size"; }; /** - * Spherical Distortion - * @description Applies spherical distortion to an image and fills the frame with the resulting image + * RecallParameter + * @description Request model for updating recallable parameters. */ - SphericalDistortionInvocation: { + RecallParameter: { /** - * @description Optional metadata to be saved with the image - * @default null + * Positive Prompt + * @description Positive prompt text */ - metadata?: components["schemas"]["MetadataField"] | null; + positive_prompt?: string | null; /** - * @description The board to save the image to - * @default null + * Negative Prompt + * @description Negative prompt text */ - board?: components["schemas"]["BoardField"] | null; + negative_prompt?: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Model + * @description Main model name/identifier */ - id: string; + model?: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Refiner Model + * @description Refiner model name/identifier */ - is_intermediate?: boolean; + refiner_model?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Vae Model + * @description VAE model name/identifier */ - use_cache?: boolean; + vae_model?: string | null; /** - * @description The image to distort - * @default null + * Scheduler + * @description Scheduler name */ - image?: components["schemas"]["ImageField"] | null; + scheduler?: string | null; /** - * K1 - * @description k1 - * @default 0.3 + * Steps + * @description Number of generation steps */ - k1?: number; + steps?: number | null; /** - * K2 - * @description k2 - * @default 0.1 + * Refiner Steps + * @description Number of refiner steps */ - k2?: number; + refiner_steps?: number | null; /** - * P1 - * @description p1 - * @default 0 + * Cfg Scale + * @description CFG scale for guidance */ - p1?: number; + cfg_scale?: number | null; /** - * P2 - * @description p2 - * @default 0 + * Cfg Rescale Multiplier + * @description CFG rescale multiplier */ - p2?: number; + cfg_rescale_multiplier?: number | null; /** - * type - * @default spherical_distortion - * @constant + * Refiner Cfg Scale + * @description Refiner CFG scale */ - type: "spherical_distortion"; - }; - /** StarredImagesResult */ - StarredImagesResult: { + refiner_cfg_scale?: number | null; /** - * Affected Boards - * @description The ids of boards affected by the delete operation + * Guidance + * @description Guidance scale */ - affected_boards: string[]; + guidance?: number | null; /** - * Starred Images - * @description The names of the images that were starred + * Width + * @description Image width in pixels */ - starred_images: string[]; - }; - /** StarterModel */ - StarterModel: { - /** Description */ - description: string; - /** Source */ - source: string; - /** Name */ - name: string; - base: components["schemas"]["BaseModelType"]; - type: components["schemas"]["ModelType"]; - format?: components["schemas"]["ModelFormat"] | null; + width?: number | null; /** - * Is Installed - * @default false + * Height + * @description Image height in pixels */ - is_installed?: boolean; + height?: number | null; /** - * Previous Names - * @default [] + * Seed + * @description Random seed */ - previous_names?: string[]; - /** Dependencies */ - dependencies?: components["schemas"]["StarterModelWithoutDependencies"][] | null; - }; - /** StarterModelBundle */ - StarterModelBundle: { - /** Name */ - name: string; - /** Models */ - models: components["schemas"]["StarterModel"][]; - }; - /** StarterModelResponse */ - StarterModelResponse: { - /** Starter Models */ - starter_models: components["schemas"]["StarterModel"][]; - /** Starter Bundles */ - starter_bundles: { - [key: string]: components["schemas"]["StarterModelBundle"]; - }; - }; - /** StarterModelWithoutDependencies */ - StarterModelWithoutDependencies: { - /** Description */ - description: string; - /** Source */ - source: string; - /** Name */ - name: string; - base: components["schemas"]["BaseModelType"]; - type: components["schemas"]["ModelType"]; - format?: components["schemas"]["ModelFormat"] | null; + seed?: number | null; /** - * Is Installed - * @default false + * Denoise Strength + * @description Denoising strength */ - is_installed?: boolean; + denoise_strength?: number | null; /** - * Previous Names - * @default [] + * Refiner Denoise Start + * @description Refiner denoising start */ - previous_names?: string[]; - }; - /** - * Store Flux Conditioning - * @description Stores a FLUX Conditioning object (CLIP and T5 embeddings) into an SQLite database. - * Returns a unique identifier for retrieval. - * Includes database size management with proactive deletion and VACUUM. - */ - StoreFluxConditioningInvocation: { + refiner_denoise_start?: number | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Clip Skip + * @description CLIP skip layers */ - id: string; + clip_skip?: number | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Seamless X + * @description Enable seamless X tiling */ - is_intermediate?: boolean; + seamless_x?: boolean | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Seamless Y + * @description Enable seamless Y tiling */ - use_cache?: boolean; + seamless_y?: boolean | null; /** - * @description The FLUX Conditioning object to store. - * @default null + * Refiner Positive Aesthetic Score + * @description Refiner positive aesthetic score */ - conditioning?: components["schemas"]["FluxConditioningField"] | null; + refiner_positive_aesthetic_score?: number | null; /** - * type - * @default store_flux_conditioning - * @constant + * Refiner Negative Aesthetic Score + * @description Refiner negative aesthetic score + */ + refiner_negative_aesthetic_score?: number | null; + /** + * Loras + * @description List of LoRAs with their weights + */ + loras?: components["schemas"]["LoRARecallParameter"][] | null; + /** + * Control Layers + * @description List of control adapters (ControlNet, T2I Adapter, Control LoRA) with their settings */ - type: "store_flux_conditioning"; + control_layers?: components["schemas"]["ControlNetRecallParameter"][] | null; + /** + * Ip Adapters + * @description List of IP Adapters with their settings + */ + ip_adapters?: components["schemas"]["IPAdapterRecallParameter"][] | null; }; /** - * String2Output - * @description Base class for invocations that output two strings + * RecallParametersUpdatedEvent + * @description Event model for recall_parameters_updated */ - String2Output: { + RecallParametersUpdatedEvent: { /** - * String 1 - * @description string 1 + * Timestamp + * @description The timestamp of the event */ - string_1: string; + timestamp: number; /** - * String 2 - * @description string 2 + * Queue Id + * @description The ID of the queue */ - string_2: string; + queue_id: string; /** - * type - * @default string_2_output - * @constant + * Parameters + * @description The recall parameters that were updated */ - type: "string_2_output"; + parameters: { + [key: string]: unknown; + }; }; /** - * String Batch - * @description Create a batched generation, where the workflow is executed once for each string in the batch. + * Create Rectangle Mask + * @description Create a rectangular mask. */ - StringBatchInvocation: { + RectangleMaskInvocation: { + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38293,112 +22988,95 @@ export type components = { */ use_cache?: boolean; /** - * Batch Group - * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. - * @default None - * @enum {string} - */ - batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; - /** - * Strings - * @description The strings to batch over + * Width + * @description The width of the entire mask. * @default null */ - strings?: string[] | null; - /** - * type - * @default string_batch - * @constant - */ - type: "string_batch"; - }; - /** - * String Collection Index - * @description CollectionIndex Picks an index out of a collection with a random option - */ - StringCollectionIndexInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + width?: number | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Height + * @description The height of the entire mask. + * @default null */ - is_intermediate?: boolean; + height?: number | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * X Left + * @description The left x-coordinate of the rectangular masked region (inclusive). + * @default null */ - use_cache?: boolean; + x_left?: number | null; /** - * Random - * @description Random Index? - * @default true + * Y Top + * @description The top y-coordinate of the rectangular masked region (inclusive). + * @default null */ - random?: boolean; + y_top?: number | null; /** - * Index - * @description zero based index into collection (note index will wrap around if out of bounds) - * @default 0 + * Rectangle Width + * @description The width of the rectangular masked region. + * @default null */ - index?: number; + rectangle_width?: number | null; /** - * Collection - * @description string collection + * Rectangle Height + * @description The height of the rectangular masked region. * @default null */ - collection?: string[] | null; + rectangle_height?: number | null; /** * type - * @default string_collection_index + * @default rectangle_mask * @constant */ - type: "string_collection_index"; + type: "rectangle_mask"; }; /** - * String Collection Primitive - * @description A collection of string primitive values + * RemoteModelFile + * @description Information about a downloadable file that forms part of a model. */ - StringCollectionInvocation: { + RemoteModelFile: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Url + * Format: uri + * @description The url to download this model file */ - id: string; + url: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Path + * Format: path + * @description The path to the file, relative to the model root */ - is_intermediate?: boolean; + path: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Size + * @description The size of this file, in bytes + * @default 0 */ - use_cache?: boolean; + size?: number | null; /** - * Collection - * @description The collection of string values - * @default [] + * Sha256 + * @description SHA256 hash of this model (not always available) */ - collection?: string[]; + sha256?: string | null; + }; + /** RemoveImagesFromBoardResult */ + RemoveImagesFromBoardResult: { /** - * type - * @default string_collection - * @constant + * Affected Boards + * @description The ids of boards affected by the delete operation */ - type: "string_collection"; + affected_boards: string[]; + /** + * Removed Images + * @description The image names that were removed from their board + */ + removed_images: string[]; }; /** - * String Collection Joiner - * @description Takes a collection of strings and returns a single string with all the collections items, separated by the input delimiter. + * Resize Latents + * @description Resizes latents to explicit width/height (in pixels). Provided dimensions are floor-divided by 8. */ - StringCollectionJoinerInvocation: { + ResizeLatentsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38417,52 +23095,70 @@ export type components = { */ use_cache?: boolean; /** - * Delimiter - * @description The character to place between each string. - * @default , + * @description Latents tensor + * @default null */ - delimiter?: string; + latents?: components["schemas"]["LatentsField"] | null; /** - * Collection - * @description The string collection to join. + * Width + * @description Width of output (px) + * @default null + */ + width?: number | null; + /** + * Height + * @description Width of output (px) * @default null */ - collection?: string[] | null; + height?: number | null; + /** + * Mode + * @description Interpolation mode + * @default bilinear + * @enum {string} + */ + mode?: "nearest" | "linear" | "bilinear" | "bicubic" | "trilinear" | "area" | "nearest-exact"; /** - * Escape Delimiter - * @description Wehter we should escape the delimiter + * Antialias + * @description Whether or not to apply antialiasing (bilinear or bicubic only) * @default false */ - escape_delimiter?: boolean; + antialias?: boolean; /** * type - * @default string_collection_joiner_invocation + * @default lresize * @constant */ - type: "string_collection_joiner_invocation"; + type: "lresize"; }; /** - * StringCollectionJoinerOutput - * @description String Collection Joiner Output + * ResourceOrigin + * @description The origin of a resource (eg image). + * + * - INTERNAL: The resource was created by the application. + * - EXTERNAL: The resource was not created by the application. + * This may be a user-initiated upload, or an internal application upload (eg Canvas init image). + * @enum {string} */ - StringCollectionJoinerOutput: { + ResourceOrigin: "internal" | "external"; + /** RetryItemsResult */ + RetryItemsResult: { /** - * Result - * @description The joined string + * Queue Id + * @description The ID of the queue */ - result: string; + queue_id: string; /** - * type - * @default string_collection_joiner_output - * @constant + * Retried Item Ids + * @description The IDs of the queue items that were retried */ - type: "string_collection_joiner_output"; + retried_item_ids: number[]; }; /** - * String Collection Primitive Linked - * @description Allows creation of collection and optionally add a collection + * Round Float + * @description Rounds a float to a specified number of decimal places. */ - StringCollectionLinkedInvocation: { + RoundInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38481,46 +23177,92 @@ export type components = { */ use_cache?: boolean; /** - * Collection - * @description The collection of string values - * @default [] + * Value + * @description The float value + * @default 0 */ - collection?: string[]; + value?: number; + /** + * Decimals + * @description The number of decimal places + * @default 0 + */ + decimals?: number; /** * type - * @default string_collection_linked + * @default round_float * @constant */ - type: "string_collection_linked"; + type: "round_float"; + }; + /** SAMPoint */ + SAMPoint: { + /** + * X + * @description The x-coordinate of the point + */ + x: number; /** - * Value - * @description The string value - * @default null + * Y + * @description The y-coordinate of the point */ - value?: string | null; + y: number; + /** @description The label of the point */ + label: components["schemas"]["SAMPointLabel"]; }; /** - * StringCollectionOutput - * @description Base class for nodes that output a collection of strings + * SAMPointLabel + * @enum {integer} */ - StringCollectionOutput: { + SAMPointLabel: -1 | 0 | 1; + /** SAMPointsField */ + SAMPointsField: { /** - * Collection - * @description The output strings + * Points + * @description The points of the object */ - collection: string[]; + points: components["schemas"]["SAMPoint"][]; + }; + /** + * SD3ConditioningField + * @description A conditioning tensor primitive value + */ + SD3ConditioningField: { + /** + * Conditioning Name + * @description The name of conditioning tensor + */ + conditioning_name: string; + }; + /** + * SD3ConditioningOutput + * @description Base class for nodes that output a single SD3 conditioning tensor + */ + SD3ConditioningOutput: { + /** @description Conditioning tensor */ + conditioning: components["schemas"]["SD3ConditioningField"]; /** * type - * @default string_collection_output + * @default sd3_conditioning_output * @constant */ - type: "string_collection_output"; - }; - /** - * String Collection Toggle - * @description Allows boolean selection between two separate string collection inputs - */ - StringCollectionToggleInvocation: { + type: "sd3_conditioning_output"; + }; + /** + * Denoise - SD3 + * @description Run denoising process with a SD3 model. + */ + SD3DenoiseInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38539,88 +23281,95 @@ export type components = { */ use_cache?: boolean; /** - * Use Second - * @description Use 2nd Input - * @default false + * @description Latents tensor + * @default null */ - use_second?: boolean; + latents?: components["schemas"]["LatentsField"] | null; /** - * Col1 - * @description First String Collection Input + * @description A mask of the region to apply the denoising process to. Values of 0.0 represent the regions to be fully denoised, and 1.0 represent the regions to be preserved. * @default null */ - col1?: string[] | null; + denoise_mask?: components["schemas"]["DenoiseMaskField"] | null; + /** + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 + */ + denoising_start?: number; + /** + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 + */ + denoising_end?: number; /** - * Col2 - * @description Second String Collection Input + * Transformer + * @description SD3 model (MMDiTX) to load * @default null */ - col2?: string[] | null; + transformer?: components["schemas"]["TransformerField"] | null; /** - * type - * @default string_collection_toggle - * @constant + * @description Positive conditioning tensor + * @default null */ - type: "string_collection_toggle"; - }; - /** - * String Generator - * @description Generated a range of strings for use in a batched generation - */ - StringGenerator: { + positive_conditioning?: components["schemas"]["SD3ConditioningField"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description Negative conditioning tensor + * @default null */ - id: string; + negative_conditioning?: components["schemas"]["SD3ConditioningField"] | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 3.5 */ - is_intermediate?: boolean; + cfg_scale?: number | number[]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Width + * @description Width of the generated image. + * @default 1024 */ - use_cache?: boolean; + width?: number; /** - * Generator Type - * @description The string generator. + * Height + * @description Height of the generated image. + * @default 1024 */ - generator: components["schemas"]["StringGeneratorField"]; + height?: number; /** - * type - * @default string_generator - * @constant + * Steps + * @description Number of steps to run + * @default 10 */ - type: "string_generator"; - }; - /** StringGeneratorField */ - StringGeneratorField: Record; - /** - * StringGeneratorOutput - * @description Base class for nodes that output a collection of strings - */ - StringGeneratorOutput: { + steps?: number; /** - * Strings - * @description The generated strings + * Seed + * @description Randomness seed for reproducibility. + * @default 0 */ - strings: string[]; + seed?: number; /** * type - * @default string_generator_output + * @default sd3_denoise * @constant */ - type: "string_generator_output"; + type: "sd3_denoise"; }; /** - * String Primitive - * @description A string primitive value + * Image to Latents - SD3 + * @description Generates latents from an image. */ - StringInvocation: { + SD3ImageToLatentsInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38639,23 +23388,37 @@ export type components = { */ use_cache?: boolean; /** - * Value - * @description The string value - * @default + * @description The image to encode + * @default null */ - value?: string; + image?: components["schemas"]["ImageField"] | null; + /** + * @description VAE + * @default null + */ + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default string + * @default sd3_i2l * @constant */ - type: "string"; + type: "sd3_i2l"; }; /** - * String Join - * @description Joins string left to string right + * Latents to Image - SD3 + * @description Generates an image from latents. */ - StringJoinInvocation: { + SD3LatentsToImageInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38674,29 +23437,27 @@ export type components = { */ use_cache?: boolean; /** - * String Left - * @description String Left - * @default + * @description Latents tensor + * @default null */ - string_left?: string; + latents?: components["schemas"]["LatentsField"] | null; /** - * String Right - * @description String Right - * @default + * @description VAE + * @default null */ - string_right?: string; + vae?: components["schemas"]["VAEField"] | null; /** * type - * @default string_join + * @default sd3_l2i * @constant */ - type: "string_join"; + type: "sd3_l2i"; }; /** - * String Join Three - * @description Joins string left to string middle to string right + * Prompt - SDXL + * @description Parse prompt using compel package to conditioning. */ - StringJoinThreeInvocation: { + SDXLCompelPromptInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38715,74 +23476,76 @@ export type components = { */ use_cache?: boolean; /** - * String Left - * @description String Left + * Prompt + * @description Prompt to be parsed by Compel to create a conditioning tensor * @default */ - string_left?: string; + prompt?: string; /** - * String Middle - * @description String Middle + * Style + * @description Prompt to be parsed by Compel to create a conditioning tensor * @default */ - string_middle?: string; + style?: string; /** - * String Right - * @description String Right - * @default + * Original Width + * @default 1024 */ - string_right?: string; + original_width?: number; /** - * type - * @default string_join_three - * @constant + * Original Height + * @default 1024 */ - type: "string_join_three"; - }; - /** - * StringOutput - * @description Base class for nodes that output a single string - */ - StringOutput: { + original_height?: number; /** - * Value - * @description The output string + * Crop Top + * @default 0 */ - value: string; + crop_top?: number; /** - * type - * @default string_output - * @constant + * Crop Left + * @default 0 */ - type: "string_output"; - }; - /** - * StringPosNegOutput - * @description Base class for invocations that output a positive and negative string - */ - StringPosNegOutput: { + crop_left?: number; /** - * Positive String - * @description Positive string + * Target Width + * @default 1024 */ - positive_string: string; + target_width?: number; /** - * Negative String - * @description Negative string + * Target Height + * @default 1024 */ - negative_string: string; + target_height?: number; + /** + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip2?: components["schemas"]["CLIPField"] | null; + /** + * @description A mask defining the region that this conditioning prompt applies to. + * @default null + */ + mask?: components["schemas"]["TensorField"] | null; /** * type - * @default string_pos_neg_output + * @default sdxl_compel_prompt * @constant */ - type: "string_pos_neg_output"; + type: "sdxl_compel_prompt"; }; /** - * String Replace - * @description Replaces the search string with the replace string + * Apply LoRA Collection - SDXL + * @description Applies a collection of SDXL LoRAs to the provided UNet and CLIP models. */ - StringReplaceInvocation: { + SDXLLoRACollectionLoader: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38801,41 +23564,41 @@ export type components = { */ use_cache?: boolean; /** - * String - * @description String to work on - * @default + * LoRAs + * @description LoRA models and weights. May be a single LoRA or collection. + * @default null */ - string?: string; + loras?: components["schemas"]["LoRAField"] | components["schemas"]["LoRAField"][] | null; /** - * Search String - * @description String to search for - * @default + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - search_string?: string; + unet?: components["schemas"]["UNetField"] | null; /** - * Replace String - * @description String to replace the search - * @default + * CLIP + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - replace_string?: string; + clip?: components["schemas"]["CLIPField"] | null; /** - * Use Regex - * @description Use search string as a regex expression (non regex is case insensitive) - * @default false + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - use_regex?: boolean; + clip2?: components["schemas"]["CLIPField"] | null; /** * type - * @default string_replace + * @default sdxl_lora_collection_loader * @constant */ - type: "string_replace"; + type: "sdxl_lora_collection_loader"; }; /** - * String Split - * @description Splits string into two strings, based on the first occurance of the delimiter. The delimiter will be removed from the string + * Apply LoRA - SDXL + * @description Apply selected lora to unet and text_encoder. */ - StringSplitInvocation: { + SDXLLoRALoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38854,29 +23617,77 @@ export type components = { */ use_cache?: boolean; /** - * String - * @description String to split - * @default + * LoRA + * @description LoRA model to load + * @default null */ - string?: string; + lora?: components["schemas"]["ModelIdentifierField"] | null; /** - * Delimiter - * @description Delimiter to spilt with. blank will split on the first whitespace - * @default + * Weight + * @description The weight at which the LoRA is applied to each model + * @default 0.75 */ - delimiter?: string; + weight?: number; + /** + * UNet + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet?: components["schemas"]["UNetField"] | null; + /** + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip?: components["schemas"]["CLIPField"] | null; + /** + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip2?: components["schemas"]["CLIPField"] | null; /** * type - * @default string_split + * @default sdxl_lora_loader * @constant */ - type: "string_split"; + type: "sdxl_lora_loader"; }; /** - * String Split Negative - * @description Splits string into two strings, inside [] goes into negative string everthing else goes into positive string. Each [ and ] character is replaced with a space + * SDXLLoRALoaderOutput + * @description SDXL LoRA Loader Output */ - StringSplitNegInvocation: { + SDXLLoRALoaderOutput: { + /** + * UNet + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet: components["schemas"]["UNetField"] | null; + /** + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip: components["schemas"]["CLIPField"] | null; + /** + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip2: components["schemas"]["CLIPField"] | null; + /** + * type + * @default sdxl_lora_loader_output + * @constant + */ + type: "sdxl_lora_loader_output"; + }; + /** + * Main Model - SDXL + * @description Loads an sdxl base model, outputting its submodels. + */ + SDXLModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38895,23 +23706,54 @@ export type components = { */ use_cache?: boolean; /** - * String - * @description String to split - * @default + * @description SDXL Main model (UNet, VAE, CLIP1, CLIP2) to load + * @default null */ - string?: string; + model?: components["schemas"]["ModelIdentifierField"] | null; /** * type - * @default string_split_neg + * @default sdxl_model_loader * @constant */ - type: "string_split_neg"; + type: "sdxl_model_loader"; + }; + /** + * SDXLModelLoaderOutput + * @description SDXL base model loader output + */ + SDXLModelLoaderOutput: { + /** + * UNet + * @description UNet (scheduler, LoRAs) + */ + unet: components["schemas"]["UNetField"]; + /** + * CLIP 1 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + */ + clip: components["schemas"]["CLIPField"]; + /** + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + */ + clip2: components["schemas"]["CLIPField"]; + /** + * VAE + * @description VAE + */ + vae: components["schemas"]["VAEField"]; + /** + * type + * @default sdxl_model_loader_output + * @constant + */ + type: "sdxl_model_loader_output"; }; /** - * String to Collection Splitter - * @description Takes a delimited string and splits it into a collection. + * Prompt - SDXL Refiner + * @description Parse prompt using compel package to conditioning. */ - StringToCollectionSplitterInvocation: { + SDXLRefinerCompelPromptInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38930,35 +23772,54 @@ export type components = { */ use_cache?: boolean; /** - * Delimiter - * @description The character dividing each string. - * @default , + * Style + * @description Prompt to be parsed by Compel to create a conditioning tensor + * @default */ - delimiter?: string; + style?: string; /** - * String - * @description The string to split. - * @default null + * Original Width + * @default 1024 */ - string?: string | null; + original_width?: number; /** - * Escape Delimiter - * @description Whether we should unescape the delimiter - * @default false + * Original Height + * @default 1024 + */ + original_height?: number; + /** + * Crop Top + * @default 0 + */ + crop_top?: number; + /** + * Crop Left + * @default 0 + */ + crop_left?: number; + /** + * Aesthetic Score + * @description The aesthetic score to apply to the conditioning tensor + * @default 6 + */ + aesthetic_score?: number; + /** + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null */ - unescape_delimiter?: boolean; + clip2?: components["schemas"]["CLIPField"] | null; /** * type - * @default string_to_collection_splitter_invocation + * @default sdxl_refiner_compel_prompt * @constant */ - type: "string_to_collection_splitter_invocation"; + type: "sdxl_refiner_compel_prompt"; }; /** - * String To Float - * @description Converts a string to a float + * Refiner Model - SDXL + * @description Loads an sdxl refiner model, outputting its submodels. */ - StringToFloatInvocation: { + SDXLRefinerModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -38977,23 +23838,64 @@ export type components = { */ use_cache?: boolean; /** - * Float String - * @description string containing a float to convert + * @description SDXL Refiner Main Modde (UNet, VAE, CLIP2) to load * @default null */ - float_string?: string | null; + model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * type + * @default sdxl_refiner_model_loader + * @constant + */ + type: "sdxl_refiner_model_loader"; + }; + /** + * SDXLRefinerModelLoaderOutput + * @description SDXL refiner model loader output + */ + SDXLRefinerModelLoaderOutput: { + /** + * UNet + * @description UNet (scheduler, LoRAs) + */ + unet: components["schemas"]["UNetField"]; + /** + * CLIP 2 + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + */ + clip2: components["schemas"]["CLIPField"]; + /** + * VAE + * @description VAE + */ + vae: components["schemas"]["VAEField"]; /** * type - * @default string_to_float + * @default sdxl_refiner_model_loader_output * @constant */ - type: "string_to_float"; - }; - /** - * String To Int - * @description Converts a string to an integer - */ - StringToIntInvocation: { + type: "sdxl_refiner_model_loader_output"; + }; + /** + * SQLiteDirection + * @enum {string} + */ + SQLiteDirection: "ASC" | "DESC"; + /** + * Save Image + * @description Saves an image. Unlike an image primitive, this invocation stores a copy of the image. + */ + SaveImageInvocation: { + /** + * @description The board to save the image to + * @default null + */ + board?: components["schemas"]["BoardField"] | null; + /** + * @description Optional metadata to be saved with the image + * @default null + */ + metadata?: components["schemas"]["MetadataField"] | null; /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -39008,27 +23910,26 @@ export type components = { /** * Use Cache * @description Whether or not to use the cache - * @default true + * @default false */ use_cache?: boolean; /** - * Int String - * @description string containing an integer to convert + * @description The image to process * @default null */ - int_string?: string | null; + image?: components["schemas"]["ImageField"] | null; /** * type - * @default string_to_int + * @default save_image * @constant */ - type: "string_to_int"; + type: "save_image"; }; /** - * String To LoRA - * @description Loads a lora from a json string, outputting its submodels. + * Scale Latents + * @description Scales latents by a given factor. */ - StringToLoraInvocation: { + ScaleLatentsInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -39047,45 +23948,41 @@ export type components = { */ use_cache?: boolean; /** - * Model String - * @description string containing a Model to convert + * @description Latents tensor * @default null */ - model_string?: string | null; + latents?: components["schemas"]["LatentsField"] | null; /** - * type - * @default string_to_lora - * @constant + * Scale Factor + * @description The factor by which to scale + * @default null */ - type: "string_to_lora"; - }; - /** - * StringToLoraOutput - * @description String to Lora model output - */ - StringToLoraOutput: { + scale_factor?: number | null; /** - * Model - * @description LoRA model to load + * Mode + * @description Interpolation mode + * @default bilinear + * @enum {string} */ - model: components["schemas"]["ModelIdentifierField"]; + mode?: "nearest" | "linear" | "bilinear" | "bicubic" | "trilinear" | "area" | "nearest-exact"; /** - * Name - * @description Model Name + * Antialias + * @description Whether or not to apply antialiasing (bilinear or bicubic only) + * @default false */ - name: string; + antialias?: boolean; /** * type - * @default string_to_lora_output + * @default lscale * @constant */ - type: "string_to_lora_output"; + type: "lscale"; }; /** - * String To Main Model - * @description Loads a main model from a json string, outputting its submodels. + * Scheduler + * @description Selects a scheduler. */ - StringToMainModelInvocation: { + SchedulerInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -39104,45 +24001,45 @@ export type components = { */ use_cache?: boolean; /** - * Model String - * @description string containing a Model to convert - * @default null + * Scheduler + * @description Scheduler to use during inference + * @default euler + * @enum {string} */ - model_string?: string | null; + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** * type - * @default string_to_main_model + * @default scheduler * @constant */ - type: "string_to_main_model"; + type: "scheduler"; }; - /** - * StringToMainModelOutput - * @description String to main model output - */ - StringToMainModelOutput: { - /** - * Model - * @description Main model (UNet, VAE, CLIP) to load - */ - model: components["schemas"]["ModelIdentifierField"]; + /** SchedulerOutput */ + SchedulerOutput: { /** - * Name - * @description Model Name + * Scheduler + * @description Scheduler to use during inference + * @enum {string} */ - name: string; + scheduler: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** * type - * @default string_to_main_model_output + * @default scheduler_output * @constant */ - type: "string_to_main_model_output"; + type: "scheduler_output"; }; /** - * String To Model - * @description Loads a model from a json string, outputting its submodels. + * SchedulerPredictionType + * @description Scheduler prediction type. + * @enum {string} + */ + SchedulerPredictionType: "epsilon" | "v_prediction" | "sample"; + /** + * Main Model - SD3 + * @description Loads a SD3 base model, outputting its submodels. */ - StringToModelInvocation: { + Sd3ModelLoaderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -39160,103 +24057,81 @@ export type components = { * @default true */ use_cache?: boolean; + /** @description SD3 model (MMDiTX) to load */ + model: components["schemas"]["ModelIdentifierField"]; /** - * Model String - * @description string containing a Model to convert + * T5 Encoder + * @description T5 tokenizer and text encoder * @default null */ - model_string?: string | null; + t5_encoder_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * type - * @default string_to_model - * @constant + * CLIP L Encoder + * @description CLIP Embed loader + * @default null */ - type: "string_to_model"; - }; - /** - * StringToModelOutput - * @description String to model output - */ - StringToModelOutput: { + clip_l_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Model - * @description Model identifier + * CLIP G Encoder + * @description CLIP-G Embed loader + * @default null */ - model: components["schemas"]["ModelIdentifierField"]; + clip_g_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Name - * @description Model Name + * VAE + * @description VAE model to load + * @default null */ - name: string; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; /** * type - * @default string_to_model_output + * @default sd3_model_loader * @constant */ - type: "string_to_model_output"; + type: "sd3_model_loader"; }; /** - * String To SDXL Main Model - * @description Loads a SDXL model from a json string, outputting its submodels. + * Sd3ModelLoaderOutput + * @description SD3 base model loader output. */ - StringToSDXLModelInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + Sd3ModelLoaderOutput: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Transformer + * @description Transformer */ - use_cache?: boolean; + transformer: components["schemas"]["TransformerField"]; /** - * Model String - * @description string containing a Model to convert - * @default null + * CLIP L + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - model_string?: string | null; + clip_l: components["schemas"]["CLIPField"]; /** - * type - * @default string_to_sdxl_model - * @constant + * CLIP G + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count */ - type: "string_to_sdxl_model"; - }; - /** - * StringToSDXLModelOutput - * @description String to SDXL main model output - */ - StringToSDXLModelOutput: { + clip_g: components["schemas"]["CLIPField"]; /** - * Model - * @description Main model (UNet, VAE, CLIP) to load + * T5 Encoder + * @description T5 tokenizer and text encoder */ - model: components["schemas"]["ModelIdentifierField"]; + t5_encoder: components["schemas"]["T5EncoderField"]; /** - * Name - * @description Model Name + * VAE + * @description VAE */ - name: string; + vae: components["schemas"]["VAEField"]; /** * type - * @default string_to_sdxl_model_output + * @default sd3_model_loader_output * @constant */ - type: "string_to_sdxl_model_output"; + type: "sd3_model_loader_output"; }; /** - * String To Scheduler - * @description Converts a string to a scheduler + * Prompt - SD3 + * @description Encodes and preps a prompt for a SD3 image. */ - StringToSchedulerInvocation: { + Sd3TextEncoderInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -39275,23 +24150,41 @@ export type components = { */ use_cache?: boolean; /** - * Scheduler String - * @description string containing a scheduler to convert + * CLIP L + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip_l?: components["schemas"]["CLIPField"] | null; + /** + * CLIP G + * @description CLIP (tokenizer, text encoder, LoRAs) and skipped layer count + * @default null + */ + clip_g?: components["schemas"]["CLIPField"] | null; + /** + * T5Encoder + * @description T5 tokenizer and text encoder + * @default null + */ + t5_encoder?: components["schemas"]["T5EncoderField"] | null; + /** + * Prompt + * @description Text prompt to encode. * @default null */ - scheduler_string?: string | null; + prompt?: string | null; /** * type - * @default string_to_scheduler + * @default sd3_text_encoder * @constant */ - type: "string_to_scheduler"; + type: "sd3_text_encoder"; }; /** - * String Toggle - * @description Allows boolean selection between two separate string inputs + * Apply Seamless - SD1.5, SDXL + * @description Applies the seamless transformation to the Model UNet and VAE. */ - StringToggleInvocation: { + SeamlessModeInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -39310,117 +24203,65 @@ export type components = { */ use_cache?: boolean; /** - * Use Second - * @description Use 2nd Input - * @default false - */ - use_second?: boolean; - /** - * String 1 - * @description First String Input + * UNet + * @description UNet (scheduler, LoRAs) * @default null */ - str1?: string | null; + unet?: components["schemas"]["UNetField"] | null; /** - * String 2 - * @description Second String Input + * VAE + * @description VAE model to load * @default null */ - str2?: string | null; - /** - * type - * @default string_toggle - * @constant - */ - type: "string_toggle"; - }; - /** - * Strings To CSV - * @description Strings To CSV converts a a list of Strings into a CSV - */ - StringsToCSVInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; - /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false - */ - is_intermediate?: boolean; + vae?: components["schemas"]["VAEField"] | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Seamless Y + * @description Specify whether Y axis is seamless + * @default true */ - use_cache?: boolean; + seamless_y?: boolean; /** - * Strings - * @description String or Collection of Strings to convert to CSV format - * @default + * Seamless X + * @description Specify whether X axis is seamless + * @default true */ - strings?: string | string[]; + seamless_x?: boolean; /** * type - * @default strings_to_csv + * @default seamless * @constant */ - type: "strings_to_csv"; + type: "seamless"; }; /** - * StylePresetField - * @description A style preset primitive field + * SeamlessModeOutput + * @description Modified Seamless Model output */ - StylePresetField: { - /** - * Style Preset Id - * @description The id of the style preset - */ - style_preset_id: string; - }; - /** StylePresetRecordWithImage */ - StylePresetRecordWithImage: { + SeamlessModeOutput: { /** - * Name - * @description The name of the style preset. - */ - name: string; - /** @description The preset data */ - preset_data: components["schemas"]["PresetData"]; - /** @description The type of style preset */ - type: components["schemas"]["PresetType"]; + * UNet + * @description UNet (scheduler, LoRAs) + * @default null + */ + unet: components["schemas"]["UNetField"] | null; /** - * Id - * @description The style preset ID. + * VAE + * @description VAE + * @default null */ - id: string; + vae: components["schemas"]["VAEField"] | null; /** - * Image - * @description The path for image + * type + * @default seamless_output + * @constant */ - image: string | null; - }; - /** - * SubModelType - * @description Submodel type. - * @enum {string} - */ - SubModelType: "unet" | "transformer" | "text_encoder" | "text_encoder_2" | "text_encoder_3" | "tokenizer" | "tokenizer_2" | "tokenizer_3" | "vae" | "vae_decoder" | "vae_encoder" | "scheduler" | "safety_checker"; - /** SubmodelDefinition */ - SubmodelDefinition: { - /** Path Or Prefix */ - path_or_prefix: string; - model_type: components["schemas"]["ModelType"]; - /** Variant */ - variant?: components["schemas"]["ModelVariantType"] | components["schemas"]["ClipVariantType"] | components["schemas"]["FluxVariantType"] | components["schemas"]["Flux2VariantType"] | components["schemas"]["ZImageVariantType"] | components["schemas"]["Qwen3VariantType"] | null; + type: "seamless_output"; }; /** - * Subtract Integers - * @description Subtracts two numbers + * Segment Anything + * @description Runs a Segment Anything Model (SAM or SAM2). */ - SubtractInvocation: { + SegmentAnythingInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -39439,330 +24280,383 @@ export type components = { */ use_cache?: boolean; /** - * A - * @description The first number - * @default 0 + * Model + * @description The Segment Anything model to use (SAM or SAM2). + * @default null */ - a?: number; + model?: ("segment-anything-base" | "segment-anything-large" | "segment-anything-huge" | "segment-anything-2-tiny" | "segment-anything-2-small" | "segment-anything-2-base" | "segment-anything-2-large") | null; /** - * B - * @description The second number - * @default 0 + * @description The image to segment. + * @default null */ - b?: number; + image?: components["schemas"]["ImageField"] | null; + /** + * Bounding Boxes + * @description The bounding boxes to prompt the model with. + * @default null + */ + bounding_boxes?: components["schemas"]["BoundingBoxField"][] | null; + /** + * Point Lists + * @description The list of point lists to prompt the model with. Each list of points represents a single object. + * @default null + */ + point_lists?: components["schemas"]["SAMPointsField"][] | null; + /** + * Apply Polygon Refinement + * @description Whether to apply polygon refinement to the masks. This will smooth the edges of the masks slightly and ensure that each mask consists of a single closed polygon (before merging). + * @default true + */ + apply_polygon_refinement?: boolean; + /** + * Mask Filter + * @description The filtering to apply to the detected masks before merging them into a final output. + * @default all + * @enum {string} + */ + mask_filter?: "all" | "largest" | "highest_box_score"; /** * type - * @default sub + * @default segment_anything * @constant */ - type: "sub"; + type: "segment_anything"; }; - /** T2IAdapterField */ - T2IAdapterField: { - /** @description The T2I-Adapter image prompt. */ - image: components["schemas"]["ImageField"]; - /** @description The T2I-Adapter model to use. */ - t2i_adapter_model: components["schemas"]["ModelIdentifierField"]; + /** SessionProcessorStatus */ + SessionProcessorStatus: { /** - * Weight - * @description The weight given to the T2I-Adapter - * @default 1 + * Is Started + * @description Whether the session processor is started */ - weight?: number | number[]; + is_started: boolean; /** - * Begin Step Percent - * @description When the T2I-Adapter is first applied (% of total steps) - * @default 0 + * Is Processing + * @description Whether a session is being processed */ - begin_step_percent?: number; + is_processing: boolean; + }; + /** + * SessionQueueAndProcessorStatus + * @description The overall status of session queue and processor + */ + SessionQueueAndProcessorStatus: { + queue: components["schemas"]["SessionQueueStatus"]; + processor: components["schemas"]["SessionProcessorStatus"]; + }; + /** SessionQueueCountsByDestination */ + SessionQueueCountsByDestination: { /** - * End Step Percent - * @description When the T2I-Adapter is last applied (% of total steps) - * @default 1 + * Queue Id + * @description The ID of the queue */ - end_step_percent?: number; + queue_id: string; /** - * Resize Mode - * @description The resize mode to use - * @default just_resize - * @enum {string} + * Destination + * @description The destination of queue items included in this status */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + destination: string; + /** + * Pending + * @description Number of queue items with status 'pending' for the destination + */ + pending: number; + /** + * In Progress + * @description Number of queue items with status 'in_progress' for the destination + */ + in_progress: number; + /** + * Completed + * @description Number of queue items with status 'complete' for the destination + */ + completed: number; + /** + * Failed + * @description Number of queue items with status 'error' for the destination + */ + failed: number; + /** + * Canceled + * @description Number of queue items with status 'canceled' for the destination + */ + canceled: number; + /** + * Total + * @description Total number of queue items for the destination + */ + total: number; }; /** - * T2I-Adapter - SD1.5, SDXL - * @description Collects T2I-Adapter info to pass to other nodes. + * SessionQueueItem + * @description Session queue item without the full graph. Used for serialization. */ - T2IAdapterInvocation: { + SessionQueueItem: { /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Item Id + * @description The identifier of the session queue item */ - id: string; + item_id: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Status + * @description The status of this queue item + * @default pending + * @enum {string} */ - is_intermediate?: boolean; + status: "pending" | "in_progress" | "completed" | "failed" | "canceled"; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Priority + * @description The priority of this queue item + * @default 0 */ - use_cache?: boolean; + priority: number; /** - * @description The IP-Adapter image prompt. - * @default null + * Batch Id + * @description The ID of the batch associated with this queue item */ - image?: components["schemas"]["ImageField"] | null; + batch_id: string; /** - * T2I-Adapter Model - * @description The T2I-Adapter model. - * @default null + * Origin + * @description The origin of this queue item. This data is used by the frontend to determine how to handle results. */ - t2i_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; + origin?: string | null; /** - * Weight - * @description The weight given to the T2I-Adapter - * @default 1 + * Destination + * @description The origin of this queue item. This data is used by the frontend to determine how to handle results */ - weight?: number | number[]; + destination?: string | null; /** - * Begin Step Percent - * @description When the T2I-Adapter is first applied (% of total steps) - * @default 0 + * Session Id + * @description The ID of the session associated with this queue item. The session doesn't exist in graph_executions until the queue item is executed. */ - begin_step_percent?: number; + session_id: string; /** - * End Step Percent - * @description When the T2I-Adapter is last applied (% of total steps) - * @default 1 + * Error Type + * @description The error type if this queue item errored */ - end_step_percent?: number; + error_type?: string | null; /** - * Resize Mode - * @description The resize mode applied to the T2I-Adapter input image so that it matches the target output size. - * @default just_resize - * @enum {string} + * Error Message + * @description The error message if this queue item errored */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + error_message?: string | null; /** - * type - * @default t2i_adapter - * @constant + * Error Traceback + * @description The error traceback if this queue item errored */ - type: "t2i_adapter"; - }; - /** - * T2I-Adapter-Linked - * @description Collects T2I-Adapter info to pass to other nodes. - */ - T2IAdapterLinkedInvocation: { + error_traceback?: string | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Created At + * @description When this queue item was created */ - id: string; + created_at: string; + /** + * Updated At + * @description When this queue item was updated + */ + updated_at: string; + /** + * Started At + * @description When this queue item was started + */ + started_at?: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Completed At + * @description When this queue item was completed */ - is_intermediate?: boolean; + completed_at?: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Queue Id + * @description The id of the queue with which this item is associated */ - use_cache?: boolean; + queue_id: string; /** - * @description The IP-Adapter image prompt. - * @default null + * User Id + * @description The id of the user who created this queue item + * @default system */ - image?: components["schemas"]["ImageField"] | null; + user_id?: string; /** - * T2I-Adapter Model - * @description The T2I-Adapter model. - * @default null + * User Display Name + * @description The display name of the user who created this queue item, if available */ - t2i_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; + user_display_name?: string | null; /** - * Weight - * @description The weight given to the T2I-Adapter - * @default 1 + * User Email + * @description The email of the user who created this queue item, if available */ - weight?: number | number[]; + user_email?: string | null; /** - * Begin Step Percent - * @description When the T2I-Adapter is first applied (% of total steps) - * @default 0 + * Field Values + * @description The field values that were used for this queue item */ - begin_step_percent?: number; + field_values?: components["schemas"]["NodeFieldValue"][] | null; /** - * End Step Percent - * @description When the T2I-Adapter is last applied (% of total steps) - * @default 1 + * Retried From Item Id + * @description The item_id of the queue item that this item was retried from */ - end_step_percent?: number; + retried_from_item_id?: number | null; + /** @description The fully-populated session to be executed */ + session: components["schemas"]["GraphExecutionState"]; + /** @description The workflow associated with this queue item */ + workflow?: components["schemas"]["WorkflowWithoutID"] | null; + }; + /** SessionQueueStatus */ + SessionQueueStatus: { /** - * Resize Mode - * @description The resize mode applied to the T2I-Adapter input image so that it matches the target output size. - * @default just_resize - * @enum {string} + * Queue Id + * @description The ID of the queue */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + queue_id: string; /** - * type - * @default t2i_adapter_linked - * @constant + * Item Id + * @description The current queue item id */ - type: "t2i_adapter_linked"; + item_id: number | null; /** - * T2I Adapter List - * @description T2I-Adapter(s) to apply - * @default null + * Batch Id + * @description The current queue item's batch id */ - t2i_adapter_list?: components["schemas"]["T2IAdapterField"] | components["schemas"]["T2IAdapterField"][] | null; - }; - /** T2IAdapterListOutput */ - T2IAdapterListOutput: { + batch_id: string | null; /** - * T2I Adapter List - * @description T2I-Adapter(s) to apply + * Session Id + * @description The current queue item's session id */ - t2i_adapter_list: components["schemas"]["T2IAdapterField"][]; + session_id: string | null; /** - * type - * @default t2i_adapter_list_output - * @constant + * Pending + * @description Number of queue items with status 'pending' */ - type: "t2i_adapter_list_output"; - }; - /** T2IAdapterMetadataField */ - T2IAdapterMetadataField: { - /** @description The control image. */ - image: components["schemas"]["ImageField"]; + pending: number; /** - * @description The control image, after processing. - * @default null + * In Progress + * @description Number of queue items with status 'in_progress' */ - processed_image?: components["schemas"]["ImageField"] | null; - /** @description The T2I-Adapter model to use. */ - t2i_adapter_model: components["schemas"]["ModelIdentifierField"]; + in_progress: number; /** - * Weight - * @description The weight given to the T2I-Adapter - * @default 1 + * Completed + * @description Number of queue items with status 'complete' */ - weight?: number | number[]; + completed: number; /** - * Begin Step Percent - * @description When the T2I-Adapter is first applied (% of total steps) - * @default 0 + * Failed + * @description Number of queue items with status 'error' */ - begin_step_percent?: number; + failed: number; /** - * End Step Percent - * @description When the T2I-Adapter is last applied (% of total steps) - * @default 1 + * Canceled + * @description Number of queue items with status 'canceled' */ - end_step_percent?: number; + canceled: number; /** - * Resize Mode - * @description The resize mode to use - * @default just_resize - * @enum {string} + * Total + * @description Total number of queue items */ - resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; - }; - /** T2IAdapterOutput */ - T2IAdapterOutput: { + total: number; /** - * T2I Adapter - * @description T2I-Adapter(s) to apply + * User Pending + * @description Number of queue items with status 'pending' for the current user */ - t2i_adapter: components["schemas"]["T2IAdapterField"]; + user_pending?: number | null; /** - * type - * @default t2i_adapter_output - * @constant + * User In Progress + * @description Number of queue items with status 'in_progress' for the current user */ - type: "t2i_adapter_output"; + user_in_progress?: number | null; }; - /** T2IAdapter_Diffusers_SD1_Config */ - T2IAdapter_Diffusers_SD1_Config: { + /** + * SetupRequest + * @description Request body for initial admin setup. + */ + SetupRequest: { /** - * Key - * @description A unique key for this model. + * Email + * @description Admin email address */ - key: string; + email: string; /** - * Hash - * @description The hash of the model file(s). + * Display Name + * @description Admin display name */ - hash: string; + display_name?: string | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Password + * @description Admin password */ - path: string; + password: string; + }; + /** + * SetupResponse + * @description Response from successful admin setup. + */ + SetupResponse: { /** - * File Size - * @description The size of the model in bytes. + * Success + * @description Whether setup was successful */ - file_size: number; + success: boolean; + /** @description Created admin user information */ + user: components["schemas"]["UserDTO"]; + }; + /** + * SetupStatusResponse + * @description Response for setup status check. + */ + SetupStatusResponse: { /** - * Name - * @description Name of the model. + * Setup Required + * @description Whether initial setup is required */ - name: string; + setup_required: boolean; /** - * Description - * @description Model description + * Multiuser Enabled + * @description Whether multiuser mode is enabled */ - description: string | null; + multiuser_enabled: boolean; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Strict Password Checking + * @description Whether strict password requirements are enforced */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + strict_password_checking: boolean; + }; + /** + * Show Image + * @description Displays a provided image using the OS image viewer, and passes it forward in the pipeline. + */ + ShowImageInvocation: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source_api_response: string | null; + id: string; /** - * Cover Image - * @description Url for image to preview model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cover_image: string | null; + is_intermediate?: boolean; /** - * Format - * @default diffusers - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + use_cache?: boolean; /** - * Type - * @default t2i_adapter - * @constant + * @description The image to show + * @default null */ - type: "t2i_adapter"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + image?: components["schemas"]["ImageField"] | null; /** - * Base - * @default sd-1 + * type + * @default show_image * @constant */ - base: "sd-1"; + type: "show_image"; }; - /** T2IAdapter_Diffusers_SDXL_Config */ - T2IAdapter_Diffusers_SDXL_Config: { + /** + * SigLIP_Diffusers_Config + * @description Model config for SigLIP. + */ + SigLIP_Diffusers_Config: { /** * Key * @description A unique key for this model. @@ -39820,112 +24714,151 @@ export type components = { repo_variant: components["schemas"]["ModelRepoVariant"]; /** * Type - * @default t2i_adapter + * @default siglip * @constant */ - type: "t2i_adapter"; - default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; + type: "siglip"; /** * Base - * @default sdxl + * @default any * @constant */ - base: "sdxl"; - }; - /** T5EncoderField */ - T5EncoderField: { - /** @description Info to load tokenizer submodel */ - tokenizer: components["schemas"]["ModelIdentifierField"]; - /** @description Info to load text_encoder submodel */ - text_encoder: components["schemas"]["ModelIdentifierField"]; + base: "any"; /** - * Loras - * @description LoRAs to apply on model loading + * Cpu Only + * @description Whether this model should run on CPU only */ - loras: components["schemas"]["LoRAField"][]; + cpu_only: boolean | null; }; /** - * T5Encoder_BnBLLMint8_Config - * @description Configuration for T5 Encoder models quantized by bitsandbytes' LLM.int8. + * Image-to-Image (Autoscale) + * @description Run any spandrel image-to-image model (https://github.com/chaiNNer-org/spandrel) until the target scale is reached. */ - T5Encoder_BnBLLMint8_Config: { + SpandrelImageToImageAutoscaleInvocation: { /** - * Key - * @description A unique key for this model. + * @description The board to save the image to + * @default null */ - key: string; + board?: components["schemas"]["BoardField"] | null; /** - * Hash - * @description The hash of the model file(s). + * @description Optional metadata to be saved with the image + * @default null */ - hash: string; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - path: string; + id: string; /** - * File Size - * @description The size of the model in bytes. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - file_size: number; + is_intermediate?: boolean; /** - * Name - * @description Name of the model. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - name: string; + use_cache?: boolean; /** - * Description - * @description Model description + * @description The input image + * @default null */ - description: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Image-to-Image Model + * @description Image-to-Image model + * @default null */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + image_to_image_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Tile Size + * @description The tile size for tiled image-to-image. Set to 0 to disable tiling. + * @default 512 */ - source_api_response: string | null; + tile_size?: number; /** - * Cover Image - * @description Url for image to preview model + * type + * @default spandrel_image_to_image_autoscale + * @constant */ - cover_image: string | null; + type: "spandrel_image_to_image_autoscale"; /** - * Base - * @default any - * @constant + * Scale + * @description The final scale of the output image. If the model does not upscale the image, this will be ignored. + * @default 4 */ - base: "any"; + scale?: number; /** - * Type - * @default t5_encoder - * @constant + * Fit To Multiple Of 8 + * @description If true, the output image will be resized to the nearest multiple of 8 in both dimensions. + * @default false */ - type: "t5_encoder"; + fit_to_multiple_of_8?: boolean; + }; + /** + * Image-to-Image + * @description Run any spandrel image-to-image model (https://github.com/chaiNNer-org/spandrel). + */ + SpandrelImageToImageInvocation: { /** - * Format - * @default bnb_quantized_int8b - * @constant + * @description The board to save the image to + * @default null */ - format: "bnb_quantized_int8b"; + board?: components["schemas"]["BoardField"] | null; /** - * Cpu Only - * @description Whether this model should run on CPU only + * @description Optional metadata to be saved with the image + * @default null */ - cpu_only: boolean | null; + metadata?: components["schemas"]["MetadataField"] | null; + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The input image + * @default null + */ + image?: components["schemas"]["ImageField"] | null; + /** + * Image-to-Image Model + * @description Image-to-Image model + * @default null + */ + image_to_image_model?: components["schemas"]["ModelIdentifierField"] | null; + /** + * Tile Size + * @description The tile size for tiled image-to-image. Set to 0 to disable tiling. + * @default 512 + */ + tile_size?: number; + /** + * type + * @default spandrel_image_to_image + * @constant + */ + type: "spandrel_image_to_image"; }; /** - * T5Encoder_T5Encoder_Config - * @description Configuration for T5 Encoder models in a bespoke, diffusers-like format. The model weights are expected to be in - * a folder called text_encoder_2 inside the model directory, with a config file named model.safetensors.index.json. + * Spandrel_Checkpoint_Config + * @description Model config for Spandrel Image to Image models. */ - T5Encoder_T5Encoder_Config: { + Spandrel_Checkpoint_Config: { /** * Key * @description A unique key for this model. @@ -39981,471 +24914,604 @@ export type components = { base: "any"; /** * Type - * @default t5_encoder + * @default spandrel_image_to_image * @constant */ - type: "t5_encoder"; + type: "spandrel_image_to_image"; /** * Format - * @default t5_encoder + * @default checkpoint * @constant */ - format: "t5_encoder"; - /** - * Cpu Only - * @description Whether this model should run on CPU only - */ - cpu_only: boolean | null; - }; - /** TBLR */ - TBLR: { - /** Top */ - top: number; - /** Bottom */ - bottom: number; - /** Left */ - left: number; - /** Right */ - right: number; + format: "checkpoint"; }; - /** TI_File_SD1_Config */ - TI_File_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; + /** StarredImagesResult */ + StarredImagesResult: { /** - * Hash - * @description The hash of the model file(s). + * Affected Boards + * @description The ids of boards affected by the delete operation */ - hash: string; + affected_boards: string[]; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Starred Images + * @description The names of the images that were starred */ - path: string; + starred_images: string[]; + }; + /** StarterModel */ + StarterModel: { + /** Description */ + description: string; + /** Source */ + source: string; + /** Name */ + name: string; + base: components["schemas"]["BaseModelType"]; + type: components["schemas"]["ModelType"]; + format?: components["schemas"]["ModelFormat"] | null; /** - * File Size - * @description The size of the model in bytes. + * Is Installed + * @default false */ - file_size: number; + is_installed?: boolean; /** - * Name - * @description Name of the model. + * Previous Names + * @default [] */ + previous_names?: string[]; + /** Dependencies */ + dependencies?: components["schemas"]["StarterModelWithoutDependencies"][] | null; + }; + /** StarterModelBundle */ + StarterModelBundle: { + /** Name */ + name: string; + /** Models */ + models: components["schemas"]["StarterModel"][]; + }; + /** StarterModelResponse */ + StarterModelResponse: { + /** Starter Models */ + starter_models: components["schemas"]["StarterModel"][]; + /** Starter Bundles */ + starter_bundles: { + [key: string]: components["schemas"]["StarterModelBundle"]; + }; + }; + /** StarterModelWithoutDependencies */ + StarterModelWithoutDependencies: { + /** Description */ + description: string; + /** Source */ + source: string; + /** Name */ name: string; + base: components["schemas"]["BaseModelType"]; + type: components["schemas"]["ModelType"]; + format?: components["schemas"]["ModelFormat"] | null; /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Installed + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_installed?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Previous Names + * @default [] */ - source_api_response: string | null; + previous_names?: string[]; + }; + /** + * String2Output + * @description Base class for invocations that output two strings + */ + String2Output: { /** - * Cover Image - * @description Url for image to preview model + * String 1 + * @description string 1 */ - cover_image: string | null; + string_1: string; /** - * Type - * @default embedding - * @constant + * String 2 + * @description string 2 */ - type: "embedding"; + string_2: string; /** - * Format - * @default embedding_file + * type + * @default string_2_output * @constant */ - format: "embedding_file"; + type: "string_2_output"; + }; + /** + * String Batch + * @description Create a batched generation, where the workflow is executed once for each string in the batch. + */ + StringBatchInvocation: { /** - * Base - * @default sd-1 - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - base: "sd-1"; - }; - /** TI_File_SD2_Config */ - TI_File_SD2_Config: { + id: string; /** - * Key - * @description A unique key for this model. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - key: string; + is_intermediate?: boolean; /** - * Hash - * @description The hash of the model file(s). + * Use Cache + * @description Whether or not to use the cache + * @default true */ - hash: string; + use_cache?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Batch Group + * @description The ID of this batch node's group. If provided, all batch nodes in with the same ID will be 'zipped' before execution, and all nodes' collections must be of the same size. + * @default None + * @enum {string} */ - path: string; + batch_group_id?: "None" | "Group 1" | "Group 2" | "Group 3" | "Group 4" | "Group 5"; /** - * File Size - * @description The size of the model in bytes. + * Strings + * @description The strings to batch over + * @default null */ - file_size: number; + strings?: string[] | null; /** - * Name - * @description Name of the model. + * type + * @default string_batch + * @constant */ - name: string; + type: "string_batch"; + }; + /** + * String Collection Primitive + * @description A collection of string primitive values + */ + StringCollectionInvocation: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Collection + * @description The collection of string values + * @default [] */ - cover_image: string | null; + collection?: string[]; /** - * Type - * @default embedding + * type + * @default string_collection * @constant */ - type: "embedding"; + type: "string_collection"; + }; + /** + * StringCollectionOutput + * @description Base class for nodes that output a collection of strings + */ + StringCollectionOutput: { /** - * Format - * @default embedding_file - * @constant + * Collection + * @description The output strings */ - format: "embedding_file"; + collection: string[]; /** - * Base - * @default sd-2 + * type + * @default string_collection_output * @constant */ - base: "sd-2"; + type: "string_collection_output"; }; - /** TI_File_SDXL_Config */ - TI_File_SDXL_Config: { + /** + * String Generator + * @description Generated a range of strings for use in a batched generation + */ + StringGenerator: { /** - * Key - * @description A unique key for this model. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - key: string; + id: string; /** - * Hash - * @description The hash of the model file(s). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - hash: string; + is_intermediate?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - path: string; + use_cache?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Generator Type + * @description The string generator. */ - file_size: number; + generator: components["schemas"]["StringGeneratorField"]; /** - * Name - * @description Name of the model. + * type + * @default string_generator + * @constant */ - name: string; + type: "string_generator"; + }; + /** StringGeneratorField */ + StringGeneratorField: Record; + /** + * StringGeneratorOutput + * @description Base class for nodes that output a collection of strings + */ + StringGeneratorOutput: { /** - * Description - * @description Model description + * Strings + * @description The generated strings */ - description: string | null; + strings: string[]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default string_generator_output + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "string_generator_output"; + }; + /** + * String Primitive + * @description A string primitive value + */ + StringInvocation: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source_api_response: string | null; + id: string; /** - * Cover Image - * @description Url for image to preview model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Type - * @default embedding - * @constant + * Value + * @description The string value + * @default */ - type: "embedding"; + value?: string; /** - * Format - * @default embedding_file + * type + * @default string * @constant */ - format: "embedding_file"; + type: "string"; + }; + /** + * String Join + * @description Joins string left to string right + */ + StringJoinInvocation: { /** - * Base - * @default sdxl - * @constant + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - base: "sdxl"; - }; - /** TI_Folder_SD1_Config */ - TI_Folder_SD1_Config: { + id: string; /** - * Key - * @description A unique key for this model. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - key: string; + is_intermediate?: boolean; /** - * Hash - * @description The hash of the model file(s). + * Use Cache + * @description Whether or not to use the cache + * @default true */ - hash: string; + use_cache?: boolean; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * String Left + * @description String Left + * @default */ - path: string; + string_left?: string; /** - * File Size - * @description The size of the model in bytes. + * String Right + * @description String Right + * @default */ - file_size: number; + string_right?: string; /** - * Name - * @description Name of the model. + * type + * @default string_join + * @constant */ - name: string; + type: "string_join"; + }; + /** + * String Join Three + * @description Joins string left to string middle to string right + */ + StringJoinThreeInvocation: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * String Left + * @description String Left + * @default */ - cover_image: string | null; + string_left?: string; /** - * Type - * @default embedding - * @constant + * String Middle + * @description String Middle + * @default */ - type: "embedding"; + string_middle?: string; /** - * Format - * @default embedding_folder - * @constant + * String Right + * @description String Right + * @default */ - format: "embedding_folder"; + string_right?: string; /** - * Base - * @default sd-1 + * type + * @default string_join_three * @constant */ - base: "sd-1"; + type: "string_join_three"; }; - /** TI_Folder_SD2_Config */ - TI_Folder_SD2_Config: { + /** + * StringOutput + * @description Base class for nodes that output a single string + */ + StringOutput: { /** - * Key - * @description A unique key for this model. + * Value + * @description The output string */ - key: string; + value: string; /** - * Hash - * @description The hash of the model file(s). + * type + * @default string_output + * @constant */ - hash: string; + type: "string_output"; + }; + /** + * StringPosNegOutput + * @description Base class for invocations that output a positive and negative string + */ + StringPosNegOutput: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Positive String + * @description Positive string */ - path: string; + positive_string: string; /** - * File Size - * @description The size of the model in bytes. + * Negative String + * @description Negative string */ - file_size: number; + negative_string: string; /** - * Name - * @description Name of the model. + * type + * @default string_pos_neg_output + * @constant */ - name: string; + type: "string_pos_neg_output"; + }; + /** + * String Replace + * @description Replaces the search string with the replace string + */ + StringReplaceInvocation: { /** - * Description - * @description Model description + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - description: string | null; + id: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + is_intermediate?: boolean; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - source_api_response: string | null; + use_cache?: boolean; /** - * Cover Image - * @description Url for image to preview model + * String + * @description String to work on + * @default */ - cover_image: string | null; + string?: string; /** - * Type - * @default embedding - * @constant + * Search String + * @description String to search for + * @default */ - type: "embedding"; + search_string?: string; /** - * Format - * @default embedding_folder - * @constant + * Replace String + * @description String to replace the search + * @default */ - format: "embedding_folder"; + replace_string?: string; /** - * Base - * @default sd-2 - * @constant + * Use Regex + * @description Use search string as a regex expression (non regex is case insensitive) + * @default false */ - base: "sd-2"; - }; - /** TI_Folder_SDXL_Config */ - TI_Folder_SDXL_Config: { + use_regex?: boolean; /** - * Key - * @description A unique key for this model. + * type + * @default string_replace + * @constant */ - key: string; + type: "string_replace"; + }; + /** + * String Split + * @description Splits string into two strings, based on the first occurance of the delimiter. The delimiter will be removed from the string + */ + StringSplitInvocation: { /** - * Hash - * @description The hash of the model file(s). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - hash: string; + id: string; /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - path: string; + is_intermediate?: boolean; /** - * File Size - * @description The size of the model in bytes. + * Use Cache + * @description Whether or not to use the cache + * @default true */ - file_size: number; + use_cache?: boolean; /** - * Name - * @description Name of the model. + * String + * @description String to split + * @default */ - name: string; + string?: string; /** - * Description - * @description Model description + * Delimiter + * @description Delimiter to spilt with. blank will split on the first whitespace + * @default */ - description: string | null; + delimiter?: string; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * type + * @default string_split + * @constant */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + type: "string_split"; + }; + /** + * String Split Negative + * @description Splits string into two strings, inside [] goes into negative string everthing else goes into positive string. Each [ and ] character is replaced with a space + */ + StringSplitNegInvocation: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source_api_response: string | null; + id: string; /** - * Cover Image - * @description Url for image to preview model + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - cover_image: string | null; + is_intermediate?: boolean; /** - * Type - * @default embedding - * @constant + * Use Cache + * @description Whether or not to use the cache + * @default true */ - type: "embedding"; + use_cache?: boolean; /** - * Format - * @default embedding_folder - * @constant + * String + * @description String to split + * @default */ - format: "embedding_folder"; + string?: string; /** - * Base - * @default sdxl + * type + * @default string_split_neg * @constant */ - base: "sdxl"; + type: "string_split_neg"; }; /** - * TensorField - * @description A tensor primitive field. + * StylePresetField + * @description A style preset primitive field */ - TensorField: { + StylePresetField: { /** - * Tensor Name - * @description The name of a tensor. + * Style Preset Id + * @description The id of the style preset */ - tensor_name: string; + style_preset_id: string; }; - /** - * Text Mask - * @description Creates a 2D rendering of a text mask from a given font. - * - * Create a white on black (or black on white) text image for use with controlnets or further processing in other nodes. Specify any TTF/OTF font file available to Invoke and control parameters to resize, rotate, and reposition the text. - * - * Currently this only generates one line of text, but it can be layered with other images using the Image Compositor node or any other such tool. - */ - TextMaskInvocation: { + /** StylePresetRecordWithImage */ + StylePresetRecordWithImage: { /** - * @description The board to save the image to - * @default null + * Name + * @description The name of the style preset. */ - board?: components["schemas"]["BoardField"] | null; + name: string; + /** @description The preset data */ + preset_data: components["schemas"]["PresetData"]; + /** @description The type of style preset */ + type: components["schemas"]["PresetType"]; /** - * @description Optional metadata to be saved with the image - * @default null + * Id + * @description The style preset ID. */ - metadata?: components["schemas"]["MetadataField"] | null; + id: string; + /** + * Image + * @description The path for image + */ + image: string | null; + }; + /** + * SubModelType + * @description Submodel type. + * @enum {string} + */ + SubModelType: "unet" | "transformer" | "text_encoder" | "text_encoder_2" | "text_encoder_3" | "tokenizer" | "tokenizer_2" | "tokenizer_3" | "vae" | "vae_decoder" | "vae_encoder" | "scheduler" | "safety_checker"; + /** SubmodelDefinition */ + SubmodelDefinition: { + /** Path Or Prefix */ + path_or_prefix: string; + model_type: components["schemas"]["ModelType"]; + /** Variant */ + variant?: components["schemas"]["ModelVariantType"] | components["schemas"]["ClipVariantType"] | components["schemas"]["FluxVariantType"] | components["schemas"]["Flux2VariantType"] | components["schemas"]["ZImageVariantType"] | components["schemas"]["Qwen3VariantType"] | null; + }; + /** + * Subtract Integers + * @description Subtracts two numbers + */ + SubtractInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -40464,83 +25530,61 @@ export type components = { */ use_cache?: boolean; /** - * Width - * @description The width of the desired mask - * @default 512 - */ - width?: number; - /** - * Height - * @description The height of the desired mask - * @default 512 + * A + * @description The first number + * @default 0 */ - height?: number; + a?: number; /** - * Text - * @description The text to render - * @default + * B + * @description The second number + * @default 0 */ - text?: string; + b?: number; /** - * Font - * @description Path to a FreeType-supported TTF/OTF font file - * @default + * type + * @default sub + * @constant */ - font?: string; + type: "sub"; + }; + /** T2IAdapterField */ + T2IAdapterField: { + /** @description The T2I-Adapter image prompt. */ + image: components["schemas"]["ImageField"]; + /** @description The T2I-Adapter model to use. */ + t2i_adapter_model: components["schemas"]["ModelIdentifierField"]; /** - * Size - * @description Desired point size of text to use - * @default 64 + * Weight + * @description The weight given to the T2I-Adapter + * @default 1 */ - size?: number; + weight?: number | number[]; /** - * Angle - * @description Angle of rotation to apply to the text + * Begin Step Percent + * @description When the T2I-Adapter is first applied (% of total steps) * @default 0 */ - angle?: number; - /** - * X Offset - * @description x-offset for text rendering - * @default 24 - */ - x_offset?: number; - /** - * Y Offset - * @description y-offset for text rendering - * @default 36 - */ - y_offset?: number; + begin_step_percent?: number; /** - * Invert - * @description Whether to invert color of the output - * @default false + * End Step Percent + * @description When the T2I-Adapter is last applied (% of total steps) + * @default 1 */ - invert?: boolean; + end_step_percent?: number; /** - * type - * @default text_mask - * @constant + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} */ - type: "text_mask"; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; }; /** - * Text to Mask Advanced (Clipseg) - * @description Uses the Clipseg model to generate an image mask from a text prompt. - * - * Output up to four prompt masks combined with logical "and", logical "or", or as separate channels of an RGBA image. + * T2I-Adapter - SD1.5, SDXL + * @description Collects T2I-Adapter info to pass to other nodes. */ - TextToMaskClipsegAdvancedInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + T2IAdapterInvocation: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -40559,817 +25603,759 @@ export type components = { */ use_cache?: boolean; /** - * @description The image from which to create a mask + * @description The IP-Adapter image prompt. * @default null */ image?: components["schemas"]["ImageField"] | null; /** - * Invert Output - * @description Off: white on black / On: black on white - * @default true - */ - invert_output?: boolean; - /** - * Prompt 1 - * @description First prompt with which to create a mask + * T2I-Adapter Model + * @description The T2I-Adapter model. * @default null */ - prompt_1?: string | null; + t2i_adapter_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Prompt 2 - * @description Second prompt with which to create a mask (optional) - * @default null + * Weight + * @description The weight given to the T2I-Adapter + * @default 1 */ - prompt_2?: string | null; + weight?: number | number[]; /** - * Prompt 3 - * @description Third prompt with which to create a mask (optional) - * @default null + * Begin Step Percent + * @description When the T2I-Adapter is first applied (% of total steps) + * @default 0 */ - prompt_3?: string | null; + begin_step_percent?: number; /** - * Prompt 4 - * @description Fourth prompt with which to create a mask (optional) - * @default null + * End Step Percent + * @description When the T2I-Adapter is last applied (% of total steps) + * @default 1 */ - prompt_4?: string | null; + end_step_percent?: number; /** - * Combine - * @description How to combine the results - * @default or + * Resize Mode + * @description The resize mode applied to the T2I-Adapter input image so that it matches the target output size. + * @default just_resize * @enum {string} */ - combine?: "or" | "and" | "butnot" | "none (rgba multiplex)"; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; /** - * Smoothing - * @description Radius of blur to apply before thresholding - * @default 4 + * type + * @default t2i_adapter + * @constant + */ + type: "t2i_adapter"; + }; + /** T2IAdapterMetadataField */ + T2IAdapterMetadataField: { + /** @description The control image. */ + image: components["schemas"]["ImageField"]; + /** + * @description The control image, after processing. + * @default null */ - smoothing?: number; + processed_image?: components["schemas"]["ImageField"] | null; + /** @description The T2I-Adapter model to use. */ + t2i_adapter_model: components["schemas"]["ModelIdentifierField"]; /** - * Subject Threshold - * @description Threshold above which is considered the subject + * Weight + * @description The weight given to the T2I-Adapter * @default 1 */ - subject_threshold?: number; + weight?: number | number[]; /** - * Background Threshold - * @description Threshold below which is considered the background + * Begin Step Percent + * @description When the T2I-Adapter is first applied (% of total steps) * @default 0 */ - background_threshold?: number; + begin_step_percent?: number; /** - * type - * @default txt2mask_clipseg_adv - * @constant + * End Step Percent + * @description When the T2I-Adapter is last applied (% of total steps) + * @default 1 */ - type: "txt2mask_clipseg_adv"; - }; - /** - * Text to Mask (Clipseg) - * @description Uses the Clipseg model to generate an image mask from a text prompt. - * - * Input a prompt and an image to generate a mask representing areas of the image matched by the prompt. - */ - TextToMaskClipsegInvocation: { + end_step_percent?: number; /** - * @description The board to save the image to - * @default null + * Resize Mode + * @description The resize mode to use + * @default just_resize + * @enum {string} */ - board?: components["schemas"]["BoardField"] | null; + resize_mode?: "just_resize" | "crop_resize" | "fill_resize" | "just_resize_simple"; + }; + /** T2IAdapterOutput */ + T2IAdapterOutput: { /** - * @description Optional metadata to be saved with the image - * @default null + * T2I Adapter + * @description T2I-Adapter(s) to apply */ - metadata?: components["schemas"]["MetadataField"] | null; + t2i_adapter: components["schemas"]["T2IAdapterField"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * type + * @default t2i_adapter_output + * @constant */ - id: string; + type: "t2i_adapter_output"; + }; + /** T2IAdapter_Diffusers_SD1_Config */ + T2IAdapter_Diffusers_SD1_Config: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Key + * @description A unique key for this model. */ - is_intermediate?: boolean; + key: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Hash + * @description The hash of the model file(s). */ - use_cache?: boolean; + hash: string; /** - * @description The image from which to create a mask - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - image?: components["schemas"]["ImageField"] | null; + path: string; /** - * Invert Output - * @description Off: white on black / On: black on white - * @default true + * File Size + * @description The size of the model in bytes. */ - invert_output?: boolean; + file_size: number; /** - * Prompt - * @description The prompt with which to create a mask - * @default null + * Name + * @description Name of the model. */ - prompt?: string | null; + name: string; /** - * Smoothing - * @description Radius of blur to apply before thresholding - * @default 4 + * Description + * @description Model description */ - smoothing?: number; + description: string | null; /** - * Subject Threshold - * @description Threshold above which is considered the subject - * @default 0.4 + * Source + * @description The original source of the model (path, URL or repo_id). */ - subject_threshold?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Background Threshold - * @description Threshold below which is considered the background - * @default 0.4 + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - background_threshold?: number; + source_api_response: string | null; /** - * Mask Expand Or Contract - * @description Pixels by which to grow (or shrink) mask after thresholding - * @default 0 + * Cover Image + * @description Url for image to preview model */ - mask_expand_or_contract?: number; + cover_image: string | null; /** - * Mask Blur - * @description Radius of blur to apply after thresholding - * @default 0 + * Format + * @default diffusers + * @constant */ - mask_blur?: number; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * type - * @default txt2mask_clipseg + * Type + * @default t2i_adapter * @constant */ - type: "txt2mask_clipseg"; - }; - /** - * Text Font to Image - * @description Turn Text into an image - */ - TextfontimageInvocation: { + type: "t2i_adapter"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Base + * @default sd-1 + * @constant */ - id: string; + base: "sd-1"; + }; + /** T2IAdapter_Diffusers_SDXL_Config */ + T2IAdapter_Diffusers_SDXL_Config: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Key + * @description A unique key for this model. */ - is_intermediate?: boolean; + key: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default false + * Hash + * @description The hash of the model file(s). */ - use_cache?: boolean; + hash: string; /** - * Text Input - * @description The text from which to generate an image - * @default Invoke AI + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - text_input?: string; + path: string; /** - * Text Input Second Row - * @description The second row of text to add below the first text - * @default null + * File Size + * @description The size of the model in bytes. */ - text_input_second_row?: string | null; + file_size: number; /** - * Second Row Font Size - * @description Font size for the second row of text (optional) - * @default 35 + * Name + * @description Name of the model. */ - second_row_font_size?: number | null; + name: string; /** - * Font Url - * @description URL address of the font file to download - * @default https://www.1001fonts.com/download/font/caliban.medium.ttf + * Description + * @description Model description */ - font_url?: string | null; + description: string | null; /** - * Local Font Path - * @description Local font file path (overrides font_url) - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - local_font_path?: string | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Local Font - * @description Name of the local font file to use from the font_cache folder - * @default None - * @constant + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - local_font?: "None"; + source_api_response: string | null; /** - * Image Width - * @description Width of the output image - * @default 1024 + * Cover Image + * @description Url for image to preview model */ - image_width?: number; + cover_image: string | null; /** - * Image Height - * @description Height of the output image - * @default 512 + * Format + * @default diffusers + * @constant */ - image_height?: number; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Padding - * @description Padding around the text in pixels - * @default 100 + * Type + * @default t2i_adapter + * @constant */ - padding?: number; + type: "t2i_adapter"; + default_settings: components["schemas"]["ControlAdapterDefaultSettings"] | null; /** - * Row Gap - * @description Gap between the two rows of text in pixels - * @default 50 + * Base + * @default sdxl + * @constant */ - row_gap?: number; + base: "sdxl"; + }; + /** T5EncoderField */ + T5EncoderField: { + /** @description Info to load tokenizer submodel */ + tokenizer: components["schemas"]["ModelIdentifierField"]; + /** @description Info to load text_encoder submodel */ + text_encoder: components["schemas"]["ModelIdentifierField"]; /** - * type - * @default Text_Font_to_Image - * @constant + * Loras + * @description LoRAs to apply on model loading */ - type: "Text_Font_to_Image"; + loras: components["schemas"]["LoRAField"][]; }; /** - * Thresholding - * @description Puts out 3 masks for a source image representing highlights, midtones, and shadows + * T5Encoder_BnBLLMint8_Config + * @description Configuration for T5 Encoder models quantized by bitsandbytes' LLM.int8. */ - ThresholdingInvocation: { + T5Encoder_BnBLLMint8_Config: { /** - * @description The board to save the image to - * @default null + * Key + * @description A unique key for this model. */ - board?: components["schemas"]["BoardField"] | null; + key: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Hash + * @description The hash of the model file(s). */ - metadata?: components["schemas"]["MetadataField"] | null; + hash: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - id: string; + path: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * File Size + * @description The size of the model in bytes. */ - is_intermediate?: boolean; + file_size: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Name + * @description Name of the model. */ - use_cache?: boolean; + name: string; /** - * @description The image to add film grain to - * @default null + * Description + * @description Model description */ - image?: components["schemas"]["ImageField"] | null; + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; /** - * Highlights Point - * @description Highlight point - * @default 170 + * Cover Image + * @description Url for image to preview model */ - highlights_point?: number; + cover_image: string | null; /** - * Shadows Point - * @description Shadow point - * @default 85 + * Base + * @default any + * @constant */ - shadows_point?: number; + base: "any"; /** - * Lut Blur - * @description LUT blur - * @default 0 + * Type + * @default t5_encoder + * @constant */ - lut_blur?: number; + type: "t5_encoder"; /** - * type - * @default thresholding + * Format + * @default bnb_quantized_int8b * @constant */ - type: "thresholding"; - }; - /** - * ThresholdingOutput - * @description Thresholding output class - */ - ThresholdingOutput: { - highlights_mask: components["schemas"]["ImageField"]; - midtones_mask: components["schemas"]["ImageField"]; - shadows_mask: components["schemas"]["ImageField"]; + format: "bnb_quantized_int8b"; /** - * type - * @default thresholding_output - * @constant + * Cpu Only + * @description Whether this model should run on CPU only */ - type: "thresholding_output"; - }; - /** Tile */ - Tile: { - /** @description The coordinates of this tile relative to its parent image. */ - coords: components["schemas"]["TBLR"]; - /** @description The amount of overlap with adjacent tiles on each side of this tile. */ - overlap: components["schemas"]["TBLR"]; + cpu_only: boolean | null; }; /** - * TileSizeOutput - * @description Tile Size Output + * T5Encoder_T5Encoder_Config + * @description Configuration for T5 Encoder models in a bespoke, diffusers-like format. The model weights are expected to be in + * a folder called text_encoder_2 inside the model directory, with a config file named model.safetensors.index.json. */ - TileSizeOutput: { - /** - * Tile Width - * @description Tile Width - */ - tile_width: number; + T5Encoder_T5Encoder_Config: { /** - * Tile Height - * @description Tile Height + * Key + * @description A unique key for this model. */ - tile_height: number; + key: string; /** - * type - * @default tile_size_output - * @constant + * Hash + * @description The hash of the model file(s). */ - type: "tile_size_output"; - }; - /** - * Tile to Properties - * @description Split a Tile into its individual properties. - */ - TileToPropertiesInvocation: { + hash: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - id: string; + path: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * File Size + * @description The size of the model in bytes. */ - is_intermediate?: boolean; + file_size: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Name + * @description Name of the model. */ - use_cache?: boolean; + name: string; /** - * @description The tile to split into properties. - * @default null + * Description + * @description Model description */ - tile?: components["schemas"]["Tile"] | null; + description: string | null; /** - * type - * @default tile_to_properties - * @constant + * Source + * @description The original source of the model (path, URL or repo_id). */ - type: "tile_to_properties"; - }; - /** TileToPropertiesOutput */ - TileToPropertiesOutput: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Coords Left - * @description Left coordinate of the tile relative to its parent image. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - coords_left: number; + source_api_response: string | null; /** - * Coords Right - * @description Right coordinate of the tile relative to its parent image. + * Cover Image + * @description Url for image to preview model */ - coords_right: number; + cover_image: string | null; /** - * Coords Top - * @description Top coordinate of the tile relative to its parent image. + * Base + * @default any + * @constant */ - coords_top: number; + base: "any"; /** - * Coords Bottom - * @description Bottom coordinate of the tile relative to its parent image. + * Type + * @default t5_encoder + * @constant */ - coords_bottom: number; + type: "t5_encoder"; /** - * Width - * @description The width of the tile. Equal to coords_right - coords_left. + * Format + * @default t5_encoder + * @constant */ - width: number; + format: "t5_encoder"; /** - * Height - * @description The height of the tile. Equal to coords_bottom - coords_top. + * Cpu Only + * @description Whether this model should run on CPU only */ - height: number; + cpu_only: boolean | null; + }; + /** TBLR */ + TBLR: { + /** Top */ + top: number; + /** Bottom */ + bottom: number; + /** Left */ + left: number; + /** Right */ + right: number; + }; + /** TI_File_SD1_Config */ + TI_File_SD1_Config: { /** - * Overlap Top - * @description Overlap between this tile and its top neighbor. + * Key + * @description A unique key for this model. */ - overlap_top: number; + key: string; /** - * Overlap Bottom - * @description Overlap between this tile and its bottom neighbor. + * Hash + * @description The hash of the model file(s). */ - overlap_bottom: number; + hash: string; /** - * Overlap Left - * @description Overlap between this tile and its left neighbor. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - overlap_left: number; + path: string; /** - * Overlap Right - * @description Overlap between this tile and its right neighbor. + * File Size + * @description The size of the model in bytes. */ - overlap_right: number; + file_size: number; /** - * type - * @default tile_to_properties_output - * @constant + * Name + * @description Name of the model. */ - type: "tile_to_properties_output"; - }; - /** TileWithImage */ - TileWithImage: { - tile: components["schemas"]["Tile"]; - image: components["schemas"]["ImageField"]; - }; - /** - * Tiled Multi-Diffusion Denoise - SD1.5, SDXL - * @description Tiled Multi-Diffusion denoising. - * - * This node handles automatically tiling the input image, and is primarily intended for global refinement of images - * in tiled upscaling workflows. Future Multi-Diffusion nodes should allow the user to specify custom regions with - * different parameters for each region to harness the full power of Multi-Diffusion. - * - * This node has a similar interface to the `DenoiseLatents` node, but it has a reduced feature set (no IP-Adapter, - * T2I-Adapter, masking, etc.). - */ - TiledMultiDiffusionDenoiseLatents: { + name: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Description + * @description Model description */ - id: string; + description: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Source + * @description The original source of the model (path, URL or repo_id). */ - is_intermediate?: boolean; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - use_cache?: boolean; + source_api_response: string | null; /** - * @description Positive conditioning tensor - * @default null + * Cover Image + * @description Url for image to preview model */ - positive_conditioning?: components["schemas"]["ConditioningField"] | null; + cover_image: string | null; /** - * @description Negative conditioning tensor - * @default null + * Type + * @default embedding + * @constant */ - negative_conditioning?: components["schemas"]["ConditioningField"] | null; + type: "embedding"; /** - * @description Noise tensor - * @default null + * Format + * @default embedding_file + * @constant */ - noise?: components["schemas"]["LatentsField"] | null; + format: "embedding_file"; /** - * @description Latents tensor - * @default null + * Base + * @default sd-1 + * @constant */ - latents?: components["schemas"]["LatentsField"] | null; + base: "sd-1"; + }; + /** TI_File_SD2_Config */ + TI_File_SD2_Config: { /** - * Tile Height - * @description Height of the tiles in image space. - * @default 1024 + * Key + * @description A unique key for this model. */ - tile_height?: number; + key: string; /** - * Tile Width - * @description Width of the tiles in image space. - * @default 1024 + * Hash + * @description The hash of the model file(s). */ - tile_width?: number; + hash: string; /** - * Tile Overlap - * @description The overlap between adjacent tiles in pixel space. (Of course, tile merging is applied in latent space.) Tiles will be cropped during merging (if necessary) to ensure that they overlap by exactly this amount. - * @default 32 + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - tile_overlap?: number; + path: string; /** - * Steps - * @description Number of steps to run - * @default 18 + * File Size + * @description The size of the model in bytes. */ - steps?: number; + file_size: number; /** - * CFG Scale - * @description Classifier-Free Guidance scale - * @default 6 + * Name + * @description Name of the model. */ - cfg_scale?: number | number[]; + name: string; /** - * Denoising Start - * @description When to start denoising, expressed a percentage of total steps - * @default 0 + * Description + * @description Model description */ - denoising_start?: number; + description: string | null; /** - * Denoising End - * @description When to stop denoising, expressed a percentage of total steps - * @default 1 + * Source + * @description The original source of the model (path, URL or repo_id). */ - denoising_end?: number; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Scheduler - * @description Scheduler to use during inference - * @default euler - * @enum {string} + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; + source_api_response: string | null; /** - * UNet - * @description UNet (scheduler, LoRAs) - * @default null + * Cover Image + * @description Url for image to preview model */ - unet?: components["schemas"]["UNetField"] | null; + cover_image: string | null; /** - * CFG Rescale Multiplier - * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR - * @default 0 + * Type + * @default embedding + * @constant */ - cfg_rescale_multiplier?: number; + type: "embedding"; /** - * Control - * @default null + * Format + * @default embedding_file + * @constant */ - control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + format: "embedding_file"; /** - * type - * @default tiled_multi_diffusion_denoise_latents + * Base + * @default sd-2 * @constant */ - type: "tiled_multi_diffusion_denoise_latents"; + base: "sd-2"; }; - /** - * TilesOutput - * @description Tiles Output - */ - TilesOutput: { + /** TI_File_SDXL_Config */ + TI_File_SDXL_Config: { /** - * Tiles - * @description Tiles Collection + * Key + * @description A unique key for this model. */ - tiles: string[]; + key: string; /** - * type - * @default tiles_output - * @constant + * Hash + * @description The hash of the model file(s). */ - type: "tiles_output"; - }; - /** - * Tracery - * @description Takes a collection of json grammars and outputs an expanded string. - */ - TraceryInvocation: { + hash: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - id: string; + path: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * File Size + * @description The size of the model in bytes. */ - is_intermediate?: boolean; + file_size: number; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Name + * @description Name of the model. */ - use_cache?: boolean; + name: string; /** - * Seed - * @description A seed for this run. - * @default 42 + * Description + * @description Model description */ - seed?: number; + description: string | null; /** - * Grammars - * @description A collection of grammars. - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - grammars?: string[] | null; + source_api_response: string | null; /** - * Prompt - * @description The prompt to expand - * @default null + * Cover Image + * @description Url for image to preview model */ - prompt?: string | null; + cover_image: string | null; /** - * type - * @default tracery_invocation + * Type + * @default embedding * @constant */ - type: "tracery_invocation"; - }; - /** - * TraceryOutput - * @description Tracery Output - */ - TraceryOutput: { + type: "embedding"; /** - * Result - * @description The expanded string + * Format + * @default embedding_file + * @constant */ - result: string; + format: "embedding_file"; /** - * type - * @default tracery_output + * Base + * @default sdxl * @constant */ - type: "tracery_output"; + base: "sdxl"; }; - /** TransformerField */ - TransformerField: { - /** @description Info to load Transformer submodel */ - transformer: components["schemas"]["ModelIdentifierField"]; + /** TI_Folder_SD1_Config */ + TI_Folder_SD1_Config: { /** - * Loras - * @description LoRAs to apply on model loading + * Key + * @description A unique key for this model. */ - loras: components["schemas"]["LoRAField"][]; - }; - /** - * UIComponent - * @description The type of UI component to use for a field, used to override the default components, which are - * inferred from the field type. - * @enum {string} - */ - UIComponent: "none" | "textarea" | "slider"; - /** - * UIConfigBase - * @description Provides additional node configuration to the UI. - * This is used internally by the @invocation decorator logic. Do not use this directly. - */ - UIConfigBase: { + key: string; /** - * Tags - * @description The node's tags - * @default null + * Hash + * @description The hash of the model file(s). */ - tags: string[] | null; + hash: string; /** - * Title - * @description The node's display name - * @default null + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - title: string | null; + path: string; /** - * Category - * @description The node's category - * @default null + * File Size + * @description The size of the model in bytes. */ - category: string | null; + file_size: number; /** - * Version - * @description The node's version. Should be a valid semver string e.g. "1.0.0" or "3.8.13". + * Name + * @description Name of the model. */ - version: string; + name: string; /** - * Node Pack - * @description The node pack that this node belongs to, will be 'invokeai' for built-in nodes + * Description + * @description Model description */ - node_pack: string; + description: string | null; /** - * @description The node's classification - * @default stable + * Source + * @description The original source of the model (path, URL or repo_id). */ - classification: components["schemas"]["Classification"]; - }; - /** - * UIType - * @description Type hints for the UI for situations in which the field type is not enough to infer the correct UI type. - * - * - Model Fields - * The most common node-author-facing use will be for model fields. Internally, there is no difference - * between SD-1, SD-2 and SDXL model fields - they all use the class `MainModelField`. To ensure the - * base-model-specific UI is rendered, use e.g. `ui_type=UIType.SDXLMainModelField` to indicate that - * the field is an SDXL main model field. - * - * - Any Field - * We cannot infer the usage of `typing.Any` via schema parsing, so you *must* use `ui_type=UIType.Any` to - * indicate that the field accepts any type. Use with caution. This cannot be used on outputs. - * - * - Scheduler Field - * Special handling in the UI is needed for this field, which otherwise would be parsed as a plain enum field. - * - * - Internal Fields - * Similar to the Any Field, the `collect` and `iterate` nodes make use of `typing.Any`. To facilitate - * handling these types in the client, we use `UIType._Collection` and `UIType._CollectionItem`. These - * should not be used by node authors. - * - * - DEPRECATED Fields - * These types are deprecated and should not be used by node authors. A warning will be logged if one is - * used, and the type will be ignored. They are included here for backwards compatibility. - * @enum {string} - */ - UIType: "SchedulerField" | "AnyField" | "CollectionField" | "CollectionItemField" | "IsIntermediate" | "DEPRECATED_Boolean" | "DEPRECATED_Color" | "DEPRECATED_Conditioning" | "DEPRECATED_Control" | "DEPRECATED_Float" | "DEPRECATED_Image" | "DEPRECATED_Integer" | "DEPRECATED_Latents" | "DEPRECATED_String" | "DEPRECATED_BooleanCollection" | "DEPRECATED_ColorCollection" | "DEPRECATED_ConditioningCollection" | "DEPRECATED_ControlCollection" | "DEPRECATED_FloatCollection" | "DEPRECATED_ImageCollection" | "DEPRECATED_IntegerCollection" | "DEPRECATED_LatentsCollection" | "DEPRECATED_StringCollection" | "DEPRECATED_BooleanPolymorphic" | "DEPRECATED_ColorPolymorphic" | "DEPRECATED_ConditioningPolymorphic" | "DEPRECATED_ControlPolymorphic" | "DEPRECATED_FloatPolymorphic" | "DEPRECATED_ImagePolymorphic" | "DEPRECATED_IntegerPolymorphic" | "DEPRECATED_LatentsPolymorphic" | "DEPRECATED_StringPolymorphic" | "DEPRECATED_UNet" | "DEPRECATED_Vae" | "DEPRECATED_CLIP" | "DEPRECATED_Collection" | "DEPRECATED_CollectionItem" | "DEPRECATED_Enum" | "DEPRECATED_WorkflowField" | "DEPRECATED_BoardField" | "DEPRECATED_MetadataItem" | "DEPRECATED_MetadataItemCollection" | "DEPRECATED_MetadataItemPolymorphic" | "DEPRECATED_MetadataDict" | "DEPRECATED_MainModelField" | "DEPRECATED_CogView4MainModelField" | "DEPRECATED_FluxMainModelField" | "DEPRECATED_SD3MainModelField" | "DEPRECATED_SDXLMainModelField" | "DEPRECATED_SDXLRefinerModelField" | "DEPRECATED_ONNXModelField" | "DEPRECATED_VAEModelField" | "DEPRECATED_FluxVAEModelField" | "DEPRECATED_LoRAModelField" | "DEPRECATED_ControlNetModelField" | "DEPRECATED_IPAdapterModelField" | "DEPRECATED_T2IAdapterModelField" | "DEPRECATED_T5EncoderModelField" | "DEPRECATED_CLIPEmbedModelField" | "DEPRECATED_CLIPLEmbedModelField" | "DEPRECATED_CLIPGEmbedModelField" | "DEPRECATED_SpandrelImageToImageModelField" | "DEPRECATED_ControlLoRAModelField" | "DEPRECATED_SigLipModelField" | "DEPRECATED_FluxReduxModelField" | "DEPRECATED_LLaVAModelField" | "DEPRECATED_Imagen3ModelField" | "DEPRECATED_Imagen4ModelField" | "DEPRECATED_ChatGPT4oModelField" | "DEPRECATED_Gemini2_5ModelField" | "DEPRECATED_FluxKontextModelField" | "DEPRECATED_Veo3ModelField" | "DEPRECATED_RunwayModelField"; - /** UNetField */ - UNetField: { - /** @description Info to load unet submodel */ - unet: components["schemas"]["ModelIdentifierField"]; - /** @description Info to load scheduler submodel */ - scheduler: components["schemas"]["ModelIdentifierField"]; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Loras - * @description LoRAs to apply on model loading + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default embedding + * @constant */ - loras: components["schemas"]["LoRAField"][]; + type: "embedding"; /** - * Seamless Axes - * @description Axes("x" and "y") to which apply seamless + * Format + * @default embedding_folder + * @constant */ - seamless_axes?: string[]; + format: "embedding_folder"; /** - * @description FreeU configuration - * @default null + * Base + * @default sd-1 + * @constant */ - freeu_config?: components["schemas"]["FreeUConfig"] | null; + base: "sd-1"; }; - /** - * UNetOutput - * @description Base class for invocations that output a UNet field. - */ - UNetOutput: { + /** TI_Folder_SD2_Config */ + TI_Folder_SD2_Config: { /** - * UNet - * @description UNet (scheduler, LoRAs) + * Key + * @description A unique key for this model. */ - unet: components["schemas"]["UNetField"]; + key: string; /** - * type - * @default unet_output - * @constant + * Hash + * @description The hash of the model file(s). */ - type: "unet_output"; - }; - /** - * URLModelSource - * @description A generic URL point to a checkpoint file. - */ - URLModelSource: { + hash: string; /** - * Url - * Format: uri + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - url: string; - /** Access Token */ - access_token?: string | null; + path: string; /** - * @description discriminator enum property added by openapi-typescript - * @enum {string} + * File Size + * @description The size of the model in bytes. */ - type: "url"; - }; - /** URLRegexTokenPair */ - URLRegexTokenPair: { + file_size: number; /** - * Url Regex - * @description Regular expression to match against the URL + * Name + * @description Name of the model. */ - url_regex: string; + name: string; /** - * Token - * @description Token to use when the URL matches the regex + * Description + * @description Model description */ - token: string; + description: string | null; + /** + * Source + * @description The original source of the model (path, URL or repo_id). + */ + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; + /** + * Source Api Response + * @description The original API response from the source, as stringified JSON. + */ + source_api_response: string | null; + /** + * Cover Image + * @description Url for image to preview model + */ + cover_image: string | null; + /** + * Type + * @default embedding + * @constant + */ + type: "embedding"; + /** + * Format + * @default embedding_folder + * @constant + */ + format: "embedding_folder"; + /** + * Base + * @default sd-2 + * @constant + */ + base: "sd-2"; }; - /** - * Unknown_Config - * @description Model config for unknown models, used as a fallback when we cannot positively identify a model. - */ - Unknown_Config: { + /** TI_Folder_SDXL_Config */ + TI_Folder_SDXL_Config: { /** * Key * @description A unique key for this model. @@ -41417,40 +26403,153 @@ export type components = { * @description Url for image to preview model */ cover_image: string | null; + /** + * Type + * @default embedding + * @constant + */ + type: "embedding"; + /** + * Format + * @default embedding_folder + * @constant + */ + format: "embedding_folder"; /** * Base - * @default unknown + * @default sdxl * @constant */ - base: "unknown"; + base: "sdxl"; + }; + /** + * TensorField + * @description A tensor primitive field. + */ + TensorField: { /** - * Type - * @default unknown + * Tensor Name + * @description The name of a tensor. + */ + tensor_name: string; + }; + /** Tile */ + Tile: { + /** @description The coordinates of this tile relative to its parent image. */ + coords: components["schemas"]["TBLR"]; + /** @description The amount of overlap with adjacent tiles on each side of this tile. */ + overlap: components["schemas"]["TBLR"]; + }; + /** + * Tile to Properties + * @description Split a Tile into its individual properties. + */ + TileToPropertiesInvocation: { + /** + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + */ + id: string; + /** + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false + */ + is_intermediate?: boolean; + /** + * Use Cache + * @description Whether or not to use the cache + * @default true + */ + use_cache?: boolean; + /** + * @description The tile to split into properties. + * @default null + */ + tile?: components["schemas"]["Tile"] | null; + /** + * type + * @default tile_to_properties * @constant */ - type: "unknown"; + type: "tile_to_properties"; + }; + /** TileToPropertiesOutput */ + TileToPropertiesOutput: { /** - * Format - * @default unknown + * Coords Left + * @description Left coordinate of the tile relative to its parent image. + */ + coords_left: number; + /** + * Coords Right + * @description Right coordinate of the tile relative to its parent image. + */ + coords_right: number; + /** + * Coords Top + * @description Top coordinate of the tile relative to its parent image. + */ + coords_top: number; + /** + * Coords Bottom + * @description Bottom coordinate of the tile relative to its parent image. + */ + coords_bottom: number; + /** + * Width + * @description The width of the tile. Equal to coords_right - coords_left. + */ + width: number; + /** + * Height + * @description The height of the tile. Equal to coords_bottom - coords_top. + */ + height: number; + /** + * Overlap Top + * @description Overlap between this tile and its top neighbor. + */ + overlap_top: number; + /** + * Overlap Bottom + * @description Overlap between this tile and its bottom neighbor. + */ + overlap_bottom: number; + /** + * Overlap Left + * @description Overlap between this tile and its left neighbor. + */ + overlap_left: number; + /** + * Overlap Right + * @description Overlap between this tile and its right neighbor. + */ + overlap_right: number; + /** + * type + * @default tile_to_properties_output * @constant */ - format: "unknown"; + type: "tile_to_properties_output"; + }; + /** TileWithImage */ + TileWithImage: { + tile: components["schemas"]["Tile"]; + image: components["schemas"]["ImageField"]; }; /** - * Unsharp Mask - * @description Applies an unsharp mask filter to an image + * Tiled Multi-Diffusion Denoise - SD1.5, SDXL + * @description Tiled Multi-Diffusion denoising. + * + * This node handles automatically tiling the input image, and is primarily intended for global refinement of images + * in tiled upscaling workflows. Future Multi-Diffusion nodes should allow the user to specify custom regions with + * different parameters for each region to harness the full power of Multi-Diffusion. + * + * This node has a similar interface to the `DenoiseLatents` node, but it has a reduced feature set (no IP-Adapter, + * T2I-Adapter, masking, etc.). */ - UnsharpMaskInvocation: { - /** - * @description The board to save the image to - * @default null - */ - board?: components["schemas"]["BoardField"] | null; - /** - * @description Optional metadata to be saved with the image - * @default null - */ - metadata?: components["schemas"]["MetadataField"] | null; + TiledMultiDiffusionDenoiseLatents: { /** * Id * @description The id of this instance of an invocation. Must be unique among all instances of invocations. @@ -41469,177 +26568,258 @@ export type components = { */ use_cache?: boolean; /** - * @description The image to use + * @description Positive conditioning tensor * @default null */ - image?: components["schemas"]["ImageField"] | null; + positive_conditioning?: components["schemas"]["ConditioningField"] | null; /** - * Radius - * @description Unsharp mask radius - * @default 2 + * @description Negative conditioning tensor + * @default null */ - radius?: number; + negative_conditioning?: components["schemas"]["ConditioningField"] | null; /** - * Strength - * @description Unsharp mask strength - * @default 50 + * @description Noise tensor + * @default null */ - strength?: number; + noise?: components["schemas"]["LatentsField"] | null; /** - * type - * @default unsharp_mask - * @constant + * @description Latents tensor + * @default null */ - type: "unsharp_mask"; - }; - /** UnstarredImagesResult */ - UnstarredImagesResult: { + latents?: components["schemas"]["LatentsField"] | null; /** - * Affected Boards - * @description The ids of boards affected by the delete operation + * Tile Height + * @description Height of the tiles in image space. + * @default 1024 */ - affected_boards: string[]; + tile_height?: number; /** - * Unstarred Images - * @description The names of the images that were unstarred + * Tile Width + * @description Width of the tiles in image space. + * @default 1024 */ - unstarred_images: string[]; - }; - /** - * UserDTO - * @description User data transfer object. - */ - UserDTO: { + tile_width?: number; /** - * User Id - * @description Unique user identifier + * Tile Overlap + * @description The overlap between adjacent tiles in pixel space. (Of course, tile merging is applied in latent space.) Tiles will be cropped during merging (if necessary) to ensure that they overlap by exactly this amount. + * @default 32 */ - user_id: string; + tile_overlap?: number; /** - * Email - * @description User email address + * Steps + * @description Number of steps to run + * @default 18 */ - email: string; + steps?: number; /** - * Display Name - * @description Display name + * CFG Scale + * @description Classifier-Free Guidance scale + * @default 6 */ - display_name?: string | null; + cfg_scale?: number | number[]; /** - * Is Admin - * @description Whether user has admin privileges - * @default false + * Denoising Start + * @description When to start denoising, expressed a percentage of total steps + * @default 0 */ - is_admin?: boolean; + denoising_start?: number; /** - * Is Active - * @description Whether user account is active - * @default true + * Denoising End + * @description When to stop denoising, expressed a percentage of total steps + * @default 1 */ - is_active?: boolean; + denoising_end?: number; /** - * Created At - * Format: date-time - * @description When the user was created + * Scheduler + * @description Scheduler to use during inference + * @default euler + * @enum {string} */ - created_at: string; + scheduler?: "ddim" | "ddpm" | "deis" | "deis_k" | "lms" | "lms_k" | "pndm" | "heun" | "heun_k" | "euler" | "euler_k" | "euler_a" | "kdpm_2" | "kdpm_2_k" | "kdpm_2_a" | "kdpm_2_a_k" | "dpmpp_2s" | "dpmpp_2s_k" | "dpmpp_2m" | "dpmpp_2m_k" | "dpmpp_2m_sde" | "dpmpp_2m_sde_k" | "dpmpp_3m" | "dpmpp_3m_k" | "dpmpp_sde" | "dpmpp_sde_k" | "unipc" | "unipc_k" | "lcm" | "tcd"; /** - * Updated At - * Format: date-time - * @description When the user was last updated + * UNet + * @description UNet (scheduler, LoRAs) + * @default null */ - updated_at: string; + unet?: components["schemas"]["UNetField"] | null; /** - * Last Login At - * @description When user last logged in + * CFG Rescale Multiplier + * @description Rescale multiplier for CFG guidance, used for models trained with zero-terminal SNR + * @default 0 */ - last_login_at?: string | null; + cfg_rescale_multiplier?: number; + /** + * Control + * @default null + */ + control?: components["schemas"]["ControlField"] | components["schemas"]["ControlField"][] | null; + /** + * type + * @default tiled_multi_diffusion_denoise_latents + * @constant + */ + type: "tiled_multi_diffusion_denoise_latents"; + }; + /** TransformerField */ + TransformerField: { + /** @description Info to load Transformer submodel */ + transformer: components["schemas"]["ModelIdentifierField"]; + /** + * Loras + * @description LoRAs to apply on model loading + */ + loras: components["schemas"]["LoRAField"][]; }; /** - * UserProfileUpdateRequest - * @description Request body for a user to update their own profile. + * UIComponent + * @description The type of UI component to use for a field, used to override the default components, which are + * inferred from the field type. + * @enum {string} */ - UserProfileUpdateRequest: { + UIComponent: "none" | "textarea" | "slider"; + /** + * UIConfigBase + * @description Provides additional node configuration to the UI. + * This is used internally by the @invocation decorator logic. Do not use this directly. + */ + UIConfigBase: { /** - * Display Name - * @description New display name + * Tags + * @description The node's tags + * @default null */ - display_name?: string | null; + tags: string[] | null; /** - * Current Password - * @description Current password (required when changing password) + * Title + * @description The node's display name + * @default null */ - current_password?: string | null; + title: string | null; /** - * New Password - * @description New password + * Category + * @description The node's category + * @default null */ - new_password?: string | null; - }; - /** VAEField */ - VAEField: { - /** @description Info to load vae submodel */ - vae: components["schemas"]["ModelIdentifierField"]; + category: string | null; /** - * Seamless Axes - * @description Axes("x" and "y") to which apply seamless + * Version + * @description The node's version. Should be a valid semver string e.g. "1.0.0" or "3.8.13". */ - seamless_axes?: string[]; + version: string; + /** + * Node Pack + * @description The node pack that this node belongs to, will be 'invokeai' for built-in nodes + */ + node_pack: string; + /** + * @description The node's classification + * @default stable + */ + classification: components["schemas"]["Classification"]; }; /** - * VAE Model - SD1.5, SD2, SDXL, SD3, FLUX - * @description Loads a VAE model, outputting a VaeLoaderOutput + * UIType + * @description Type hints for the UI for situations in which the field type is not enough to infer the correct UI type. + * + * - Model Fields + * The most common node-author-facing use will be for model fields. Internally, there is no difference + * between SD-1, SD-2 and SDXL model fields - they all use the class `MainModelField`. To ensure the + * base-model-specific UI is rendered, use e.g. `ui_type=UIType.SDXLMainModelField` to indicate that + * the field is an SDXL main model field. + * + * - Any Field + * We cannot infer the usage of `typing.Any` via schema parsing, so you *must* use `ui_type=UIType.Any` to + * indicate that the field accepts any type. Use with caution. This cannot be used on outputs. + * + * - Scheduler Field + * Special handling in the UI is needed for this field, which otherwise would be parsed as a plain enum field. + * + * - Internal Fields + * Similar to the Any Field, the `collect` and `iterate` nodes make use of `typing.Any`. To facilitate + * handling these types in the client, we use `UIType._Collection` and `UIType._CollectionItem`. These + * should not be used by node authors. + * + * - DEPRECATED Fields + * These types are deprecated and should not be used by node authors. A warning will be logged if one is + * used, and the type will be ignored. They are included here for backwards compatibility. + * @enum {string} */ - VAELoaderInvocation: { - /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. - */ - id: string; + UIType: "SchedulerField" | "AnyField" | "CollectionField" | "CollectionItemField" | "IsIntermediate" | "DEPRECATED_Boolean" | "DEPRECATED_Color" | "DEPRECATED_Conditioning" | "DEPRECATED_Control" | "DEPRECATED_Float" | "DEPRECATED_Image" | "DEPRECATED_Integer" | "DEPRECATED_Latents" | "DEPRECATED_String" | "DEPRECATED_BooleanCollection" | "DEPRECATED_ColorCollection" | "DEPRECATED_ConditioningCollection" | "DEPRECATED_ControlCollection" | "DEPRECATED_FloatCollection" | "DEPRECATED_ImageCollection" | "DEPRECATED_IntegerCollection" | "DEPRECATED_LatentsCollection" | "DEPRECATED_StringCollection" | "DEPRECATED_BooleanPolymorphic" | "DEPRECATED_ColorPolymorphic" | "DEPRECATED_ConditioningPolymorphic" | "DEPRECATED_ControlPolymorphic" | "DEPRECATED_FloatPolymorphic" | "DEPRECATED_ImagePolymorphic" | "DEPRECATED_IntegerPolymorphic" | "DEPRECATED_LatentsPolymorphic" | "DEPRECATED_StringPolymorphic" | "DEPRECATED_UNet" | "DEPRECATED_Vae" | "DEPRECATED_CLIP" | "DEPRECATED_Collection" | "DEPRECATED_CollectionItem" | "DEPRECATED_Enum" | "DEPRECATED_WorkflowField" | "DEPRECATED_BoardField" | "DEPRECATED_MetadataItem" | "DEPRECATED_MetadataItemCollection" | "DEPRECATED_MetadataItemPolymorphic" | "DEPRECATED_MetadataDict" | "DEPRECATED_MainModelField" | "DEPRECATED_CogView4MainModelField" | "DEPRECATED_FluxMainModelField" | "DEPRECATED_SD3MainModelField" | "DEPRECATED_SDXLMainModelField" | "DEPRECATED_SDXLRefinerModelField" | "DEPRECATED_ONNXModelField" | "DEPRECATED_VAEModelField" | "DEPRECATED_FluxVAEModelField" | "DEPRECATED_LoRAModelField" | "DEPRECATED_ControlNetModelField" | "DEPRECATED_IPAdapterModelField" | "DEPRECATED_T2IAdapterModelField" | "DEPRECATED_T5EncoderModelField" | "DEPRECATED_CLIPEmbedModelField" | "DEPRECATED_CLIPLEmbedModelField" | "DEPRECATED_CLIPGEmbedModelField" | "DEPRECATED_SpandrelImageToImageModelField" | "DEPRECATED_ControlLoRAModelField" | "DEPRECATED_SigLipModelField" | "DEPRECATED_FluxReduxModelField" | "DEPRECATED_LLaVAModelField" | "DEPRECATED_Imagen3ModelField" | "DEPRECATED_Imagen4ModelField" | "DEPRECATED_ChatGPT4oModelField" | "DEPRECATED_Gemini2_5ModelField" | "DEPRECATED_FluxKontextModelField" | "DEPRECATED_Veo3ModelField" | "DEPRECATED_RunwayModelField"; + /** UNetField */ + UNetField: { + /** @description Info to load unet submodel */ + unet: components["schemas"]["ModelIdentifierField"]; + /** @description Info to load scheduler submodel */ + scheduler: components["schemas"]["ModelIdentifierField"]; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Loras + * @description LoRAs to apply on model loading */ - is_intermediate?: boolean; + loras: components["schemas"]["LoRAField"][]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Seamless Axes + * @description Axes("x" and "y") to which apply seamless */ - use_cache?: boolean; + seamless_axes?: string[]; /** - * VAE - * @description VAE model to load + * @description FreeU configuration * @default null */ - vae_model?: components["schemas"]["ModelIdentifierField"] | null; + freeu_config?: components["schemas"]["FreeUConfig"] | null; + }; + /** + * UNetOutput + * @description Base class for invocations that output a UNet field. + */ + UNetOutput: { + /** + * UNet + * @description UNet (scheduler, LoRAs) + */ + unet: components["schemas"]["UNetField"]; /** * type - * @default vae_loader + * @default unet_output * @constant */ - type: "vae_loader"; + type: "unet_output"; }; /** - * VAEOutput - * @description Base class for invocations that output a VAE field + * URLModelSource + * @description A generic URL point to a checkpoint file. */ - VAEOutput: { + URLModelSource: { /** - * VAE - * @description VAE + * Url + * Format: uri */ - vae: components["schemas"]["VAEField"]; + url: string; + /** Access Token */ + access_token?: string | null; /** - * type - * @default vae_output - * @constant + * @description discriminator enum property added by openapi-typescript + * @enum {string} */ - type: "vae_output"; + type: "url"; }; - /** VAE_Checkpoint_FLUX_Config */ - VAE_Checkpoint_FLUX_Config: { + /** URLRegexTokenPair */ + URLRegexTokenPair: { + /** + * Url Regex + * @description Regular expression to match against the URL + */ + url_regex: string; + /** + * Token + * @description Token to use when the URL matches the regex + */ + token: string; + }; + /** + * Unknown_Config + * @description Model config for unknown models, used as a fallback when we cannot positively identify a model. + */ + Unknown_Config: { /** * Key * @description A unique key for this model. @@ -41688,253 +26868,228 @@ export type components = { */ cover_image: string | null; /** - * Config Path - * @description Path to the config for this model, if any. + * Base + * @default unknown + * @constant */ - config_path: string | null; + base: "unknown"; /** * Type - * @default vae + * @default unknown * @constant */ - type: "vae"; + type: "unknown"; /** * Format - * @default checkpoint - * @constant - */ - format: "checkpoint"; - /** - * Base - * @default flux + * @default unknown * @constant */ - base: "flux"; + format: "unknown"; }; /** - * VAE_Checkpoint_Flux2_Config - * @description Model config for FLUX.2 VAE checkpoint models (AutoencoderKLFlux2). + * Unsharp Mask + * @description Applies an unsharp mask filter to an image */ - VAE_Checkpoint_Flux2_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; + UnsharpMaskInvocation: { /** - * Name - * @description Name of the model. + * @description The board to save the image to + * @default null */ - name: string; + board?: components["schemas"]["BoardField"] | null; /** - * Description - * @description Model description + * @description Optional metadata to be saved with the image + * @default null */ - description: string | null; + metadata?: components["schemas"]["MetadataField"] | null; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * @description The image to use + * @default null */ - config_path: string | null; + image?: components["schemas"]["ImageField"] | null; /** - * Type - * @default vae - * @constant + * Radius + * @description Unsharp mask radius + * @default 2 */ - type: "vae"; + radius?: number; /** - * Format - * @default checkpoint - * @constant + * Strength + * @description Unsharp mask strength + * @default 50 */ - format: "checkpoint"; + strength?: number; /** - * Base - * @default flux2 + * type + * @default unsharp_mask * @constant */ - base: "flux2"; + type: "unsharp_mask"; }; - /** VAE_Checkpoint_SD1_Config */ - VAE_Checkpoint_SD1_Config: { - /** - * Key - * @description A unique key for this model. - */ - key: string; - /** - * Hash - * @description The hash of the model file(s). - */ - hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; + /** UnstarredImagesResult */ + UnstarredImagesResult: { /** - * Description - * @description Model description + * Affected Boards + * @description The ids of boards affected by the delete operation */ - description: string | null; + affected_boards: string[]; /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Unstarred Images + * @description The names of the images that were unstarred */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + unstarred_images: string[]; + }; + /** + * UserDTO + * @description User data transfer object. + */ + UserDTO: { /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * User Id + * @description Unique user identifier */ - source_api_response: string | null; + user_id: string; /** - * Cover Image - * @description Url for image to preview model + * Email + * @description User email address */ - cover_image: string | null; + email: string; /** - * Config Path - * @description Path to the config for this model, if any. + * Display Name + * @description Display name */ - config_path: string | null; + display_name?: string | null; /** - * Type - * @default vae - * @constant + * Is Admin + * @description Whether user has admin privileges + * @default false */ - type: "vae"; + is_admin?: boolean; /** - * Format - * @default checkpoint - * @constant + * Is Active + * @description Whether user account is active + * @default true */ - format: "checkpoint"; + is_active?: boolean; /** - * Base - * @default sd-1 - * @constant + * Created At + * Format: date-time + * @description When the user was created */ - base: "sd-1"; - }; - /** VAE_Checkpoint_SD2_Config */ - VAE_Checkpoint_SD2_Config: { + created_at: string; /** - * Key - * @description A unique key for this model. + * Updated At + * Format: date-time + * @description When the user was last updated */ - key: string; + updated_at: string; /** - * Hash - * @description The hash of the model file(s). + * Last Login At + * @description When user last logged in */ - hash: string; + last_login_at?: string | null; + }; + /** + * UserProfileUpdateRequest + * @description Request body for a user to update their own profile. + */ + UserProfileUpdateRequest: { /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. + * Display Name + * @description New display name */ - path: string; + display_name?: string | null; /** - * File Size - * @description The size of the model in bytes. + * Current Password + * @description Current password (required when changing password) */ - file_size: number; + current_password?: string | null; /** - * Name - * @description Name of the model. + * New Password + * @description New password */ - name: string; + new_password?: string | null; + }; + /** VAEField */ + VAEField: { + /** @description Info to load vae submodel */ + vae: components["schemas"]["ModelIdentifierField"]; /** - * Description - * @description Model description + * Seamless Axes + * @description Axes("x" and "y") to which apply seamless */ - description: string | null; + seamless_axes?: string[]; + }; + /** + * VAE Model - SD1.5, SD2, SDXL, SD3, FLUX + * @description Loads a VAE model, outputting a VaeLoaderOutput + */ + VAELoaderInvocation: { /** - * Source - * @description The original source of the model (path, URL or repo_id). + * Id + * @description The id of this instance of an invocation. Must be unique among all instances of invocations. */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; + id: string; /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. + * Is Intermediate + * @description Whether or not this is an intermediate invocation. + * @default false */ - source_api_response: string | null; + is_intermediate?: boolean; /** - * Cover Image - * @description Url for image to preview model + * Use Cache + * @description Whether or not to use the cache + * @default true */ - cover_image: string | null; + use_cache?: boolean; /** - * Config Path - * @description Path to the config for this model, if any. + * VAE + * @description VAE model to load + * @default null */ - config_path: string | null; + vae_model?: components["schemas"]["ModelIdentifierField"] | null; /** - * Type - * @default vae + * type + * @default vae_loader * @constant */ - type: "vae"; + type: "vae_loader"; + }; + /** + * VAEOutput + * @description Base class for invocations that output a VAE field + */ + VAEOutput: { /** - * Format - * @default checkpoint - * @constant + * VAE + * @description VAE */ - format: "checkpoint"; + vae: components["schemas"]["VAEField"]; /** - * Base - * @default sd-2 + * type + * @default vae_output * @constant */ - base: "sd-2"; + type: "vae_output"; }; - /** VAE_Checkpoint_SDXL_Config */ - VAE_Checkpoint_SDXL_Config: { + /** VAE_Checkpoint_FLUX_Config */ + VAE_Checkpoint_FLUX_Config: { /** * Key * @description A unique key for this model. @@ -42001,16 +27156,16 @@ export type components = { format: "checkpoint"; /** * Base - * @default sdxl + * @default flux * @constant */ - base: "sdxl"; + base: "flux"; }; /** - * VAE_Diffusers_Flux2_Config - * @description Model config for FLUX.2 VAE models in diffusers format (AutoencoderKLFlux2). + * VAE_Checkpoint_Flux2_Config + * @description Model config for FLUX.2 VAE checkpoint models (AutoencoderKLFlux2). */ - VAE_Diffusers_Flux2_Config: { + VAE_Checkpoint_Flux2_Config: { /** * Key * @description A unique key for this model. @@ -42059,19 +27214,22 @@ export type components = { */ cover_image: string | null; /** - * Format - * @default diffusers - * @constant + * Config Path + * @description Path to the config for this model, if any. */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + config_path: string | null; /** * Type * @default vae * @constant */ type: "vae"; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; /** * Base * @default flux2 @@ -42079,8 +27237,8 @@ export type components = { */ base: "flux2"; }; - /** VAE_Diffusers_SD1_Config */ - VAE_Diffusers_SD1_Config: { + /** VAE_Checkpoint_SD1_Config */ + VAE_Checkpoint_SD1_Config: { /** * Key * @description A unique key for this model. @@ -42129,19 +27287,22 @@ export type components = { */ cover_image: string | null; /** - * Format - * @default diffusers - * @constant + * Config Path + * @description Path to the config for this model, if any. */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; + config_path: string | null; /** * Type * @default vae * @constant */ type: "vae"; + /** + * Format + * @default checkpoint + * @constant + */ + format: "checkpoint"; /** * Base * @default sd-1 @@ -42149,8 +27310,8 @@ export type components = { */ base: "sd-1"; }; - /** VAE_Diffusers_SDXL_Config */ - VAE_Diffusers_SDXL_Config: { + /** VAE_Checkpoint_SD2_Config */ + VAE_Checkpoint_SD2_Config: { /** * Key * @description A unique key for this model. @@ -42161,787 +27322,642 @@ export type components = { * @description The hash of the model file(s). */ hash: string; - /** - * Path - * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. - */ - path: string; - /** - * File Size - * @description The size of the model in bytes. - */ - file_size: number; - /** - * Name - * @description Name of the model. - */ - name: string; - /** - * Description - * @description Model description - */ - description: string | null; - /** - * Source - * @description The original source of the model (path, URL or repo_id). - */ - source: string; - /** @description The type of source */ - source_type: components["schemas"]["ModelSourceType"]; - /** - * Source Api Response - * @description The original API response from the source, as stringified JSON. - */ - source_api_response: string | null; - /** - * Cover Image - * @description Url for image to preview model - */ - cover_image: string | null; - /** - * Format - * @default diffusers - * @constant - */ - format: "diffusers"; - /** @default */ - repo_variant: components["schemas"]["ModelRepoVariant"]; - /** - * Type - * @default vae - * @constant - */ - type: "vae"; - /** - * Base - * @default sdxl - * @constant - */ - base: "sdxl"; - }; - /** ValidationError */ - ValidationError: { - /** Location */ - loc: (string | number)[]; - /** Message */ - msg: string; - /** Error Type */ - type: string; - }; - /** WeightedStringOutput */ - WeightedStringOutput: { - /** - * Cleaned Text - * @description The input string with weights and parentheses removed - */ - cleaned_text: string; - /** - * Phrases - * @description List of weighted phrases or words - */ - phrases: string[]; - /** - * Weights - * @description Associated weights for each phrase - */ - weights: number[]; - /** - * Positions - * @description Start positions of each phrase in the cleaned string - */ - positions: number[]; - /** - * type - * @default weighted_string_output - * @constant - */ - type: "weighted_string_output"; - }; - /** Workflow */ - Workflow: { - /** - * Name - * @description The name of the workflow. - */ - name: string; - /** - * Author - * @description The author of the workflow. + /** + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - author: string; + path: string; /** - * Description - * @description The description of the workflow. + * File Size + * @description The size of the model in bytes. */ - description: string; + file_size: number; /** - * Version - * @description The version of the workflow. + * Name + * @description Name of the model. */ - version: string; + name: string; /** - * Contact - * @description The contact of the workflow. + * Description + * @description Model description */ - contact: string; + description: string | null; /** - * Tags - * @description The tags of the workflow. + * Source + * @description The original source of the model (path, URL or repo_id). */ - tags: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Notes - * @description The notes of the workflow. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - notes: string; + source_api_response: string | null; /** - * Exposedfields - * @description The exposed fields of the workflow. + * Cover Image + * @description Url for image to preview model */ - exposedFields: components["schemas"]["ExposedField"][]; - /** @description The meta of the workflow. */ - meta: components["schemas"]["WorkflowMeta"]; + cover_image: string | null; /** - * Nodes - * @description The nodes of the workflow. + * Config Path + * @description Path to the config for this model, if any. */ - nodes: { - [key: string]: components["schemas"]["JsonValue"]; - }[]; + config_path: string | null; /** - * Edges - * @description The edges of the workflow. + * Type + * @default vae + * @constant */ - edges: { - [key: string]: components["schemas"]["JsonValue"]; - }[]; + type: "vae"; /** - * Form - * @description The form of the workflow. + * Format + * @default checkpoint + * @constant */ - form?: { - [key: string]: components["schemas"]["JsonValue"]; - } | null; + format: "checkpoint"; /** - * Id - * @description The id of the workflow. + * Base + * @default sd-2 + * @constant */ - id: string; + base: "sd-2"; }; - /** WorkflowAndGraphResponse */ - WorkflowAndGraphResponse: { + /** VAE_Checkpoint_SDXL_Config */ + VAE_Checkpoint_SDXL_Config: { /** - * Workflow - * @description The workflow used to generate the image, as stringified JSON + * Key + * @description A unique key for this model. */ - workflow: string | null; + key: string; /** - * Graph - * @description The graph used to generate the image, as stringified JSON + * Hash + * @description The hash of the model file(s). */ - graph: string | null; - }; - /** - * WorkflowCategory - * @enum {string} - */ - WorkflowCategory: "user" | "default"; - /** WorkflowMeta */ - WorkflowMeta: { + hash: string; /** - * Version - * @description The version of the workflow schema. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - version: string; - /** @description The category of the workflow (user or default). */ - category: components["schemas"]["WorkflowCategory"]; - }; - /** WorkflowRecordDTO */ - WorkflowRecordDTO: { + path: string; /** - * Workflow Id - * @description The id of the workflow. + * File Size + * @description The size of the model in bytes. */ - workflow_id: string; + file_size: number; /** * Name - * @description The name of the workflow. + * @description Name of the model. */ name: string; /** - * Created At - * @description The created timestamp of the workflow. + * Description + * @description Model description */ - created_at: string; + description: string | null; /** - * Updated At - * @description The updated timestamp of the workflow. + * Source + * @description The original source of the model (path, URL or repo_id). */ - updated_at: string; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Opened At - * @description The opened timestamp of the workflow. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - opened_at?: string | null; - /** @description The workflow. */ - workflow: components["schemas"]["Workflow"]; - }; - /** WorkflowRecordListItemWithThumbnailDTO */ - WorkflowRecordListItemWithThumbnailDTO: { + source_api_response: string | null; /** - * Workflow Id - * @description The id of the workflow. + * Cover Image + * @description Url for image to preview model */ - workflow_id: string; + cover_image: string | null; /** - * Name - * @description The name of the workflow. + * Config Path + * @description Path to the config for this model, if any. */ - name: string; + config_path: string | null; /** - * Created At - * @description The created timestamp of the workflow. + * Type + * @default vae + * @constant */ - created_at: string; + type: "vae"; /** - * Updated At - * @description The updated timestamp of the workflow. + * Format + * @default checkpoint + * @constant */ - updated_at: string; + format: "checkpoint"; /** - * Opened At - * @description The opened timestamp of the workflow. + * Base + * @default sdxl + * @constant */ - opened_at?: string | null; + base: "sdxl"; + }; + /** + * VAE_Diffusers_Flux2_Config + * @description Model config for FLUX.2 VAE models in diffusers format (AutoencoderKLFlux2). + */ + VAE_Diffusers_Flux2_Config: { /** - * Description - * @description The description of the workflow. + * Key + * @description A unique key for this model. */ - description: string; - /** @description The description of the workflow. */ - category: components["schemas"]["WorkflowCategory"]; + key: string; /** - * Tags - * @description The tags of the workflow. + * Hash + * @description The hash of the model file(s). */ - tags: string; + hash: string; /** - * Thumbnail Url - * @description The URL of the workflow thumbnail. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - thumbnail_url?: string | null; - }; - /** - * WorkflowRecordOrderBy - * @description The order by options for workflow records - * @enum {string} - */ - WorkflowRecordOrderBy: "created_at" | "updated_at" | "opened_at" | "name"; - /** WorkflowRecordWithThumbnailDTO */ - WorkflowRecordWithThumbnailDTO: { + path: string; /** - * Workflow Id - * @description The id of the workflow. + * File Size + * @description The size of the model in bytes. */ - workflow_id: string; + file_size: number; /** * Name - * @description The name of the workflow. + * @description Name of the model. */ name: string; /** - * Created At - * @description The created timestamp of the workflow. - */ - created_at: string; - /** - * Updated At - * @description The updated timestamp of the workflow. + * Description + * @description Model description */ - updated_at: string; + description: string | null; /** - * Opened At - * @description The opened timestamp of the workflow. + * Source + * @description The original source of the model (path, URL or repo_id). */ - opened_at?: string | null; - /** @description The workflow. */ - workflow: components["schemas"]["Workflow"]; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Thumbnail Url - * @description The URL of the workflow thumbnail. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - thumbnail_url?: string | null; - }; - /** WorkflowWithoutID */ - WorkflowWithoutID: { + source_api_response: string | null; /** - * Name - * @description The name of the workflow. + * Cover Image + * @description Url for image to preview model */ - name: string; + cover_image: string | null; /** - * Author - * @description The author of the workflow. + * Format + * @default diffusers + * @constant */ - author: string; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Description - * @description The description of the workflow. + * Type + * @default vae + * @constant */ - description: string; + type: "vae"; /** - * Version - * @description The version of the workflow. + * Base + * @default flux2 + * @constant */ - version: string; + base: "flux2"; + }; + /** VAE_Diffusers_SD1_Config */ + VAE_Diffusers_SD1_Config: { /** - * Contact - * @description The contact of the workflow. + * Key + * @description A unique key for this model. */ - contact: string; + key: string; /** - * Tags - * @description The tags of the workflow. + * Hash + * @description The hash of the model file(s). */ - tags: string; + hash: string; /** - * Notes - * @description The notes of the workflow. + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - notes: string; + path: string; /** - * Exposedfields - * @description The exposed fields of the workflow. + * File Size + * @description The size of the model in bytes. */ - exposedFields: components["schemas"]["ExposedField"][]; - /** @description The meta of the workflow. */ - meta: components["schemas"]["WorkflowMeta"]; + file_size: number; /** - * Nodes - * @description The nodes of the workflow. + * Name + * @description Name of the model. */ - nodes: { - [key: string]: components["schemas"]["JsonValue"]; - }[]; + name: string; /** - * Edges - * @description The edges of the workflow. + * Description + * @description Model description */ - edges: { - [key: string]: components["schemas"]["JsonValue"]; - }[]; + description: string | null; /** - * Form - * @description The form of the workflow. + * Source + * @description The original source of the model (path, URL or repo_id). */ - form?: { - [key: string]: components["schemas"]["JsonValue"]; - } | null; - }; - /** - * XY Expand - * @description Takes an XY Item and outputs the X and Y as individual strings - */ - XYExpandInvocation: { + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - id: string; + source_api_response: string | null; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Cover Image + * @description Url for image to preview model */ - is_intermediate?: boolean; + cover_image: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Format + * @default diffusers + * @constant */ - use_cache?: boolean; + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Xy Item - * @description The XY Item - * @default null + * Type + * @default vae + * @constant */ - xy_item?: string | null; + type: "vae"; /** - * type - * @default xy_expand + * Base + * @default sd-1 * @constant */ - type: "xy_expand"; + base: "sd-1"; }; - /** - * XYExpandOutput - * @description Two strings that are expanded from an XY Item - */ - XYExpandOutput: { + /** VAE_Diffusers_SDXL_Config */ + VAE_Diffusers_SDXL_Config: { /** - * X Item - * @description The X item + * Key + * @description A unique key for this model. */ - x_item: string; + key: string; /** - * Y Item - * @description The y item + * Hash + * @description The hash of the model file(s). */ - y_item: string; + hash: string; /** - * type - * @default xy_expand_output - * @constant + * Path + * @description Path to the model on the filesystem. Relative paths are relative to the Invoke root directory. */ - type: "xy_expand_output"; - }; - /** - * XYImage Collect - * @description Takes xItem, yItem and an Image and outputs it as an XYImage Item (x_item,y_item,image_name)array converted to json - */ - XYImageCollectInvocation: { + path: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * File Size + * @description The size of the model in bytes. */ - id: string; + file_size: number; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Name + * @description Name of the model. */ - is_intermediate?: boolean; + name: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Description + * @description Model description */ - use_cache?: boolean; + description: string | null; /** - * X Item - * @description The X item - * @default null + * Source + * @description The original source of the model (path, URL or repo_id). */ - x_item?: string | null; + source: string; + /** @description The type of source */ + source_type: components["schemas"]["ModelSourceType"]; /** - * Y Item - * @description The Y item - * @default null + * Source Api Response + * @description The original API response from the source, as stringified JSON. */ - y_item?: string | null; + source_api_response: string | null; /** - * @description The image to turn into grids - * @default null + * Cover Image + * @description Url for image to preview model */ - image?: components["schemas"]["ImageField"] | null; + cover_image: string | null; /** - * type - * @default xy_image_collect + * Format + * @default diffusers * @constant */ - type: "xy_image_collect"; - }; - /** - * XYImage Expand - * @description Takes an XYImage item and outputs the XItem,YItem, Image, width & height - */ - XYImageExpandInvocation: { + format: "diffusers"; + /** @default */ + repo_variant: components["schemas"]["ModelRepoVariant"]; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Type + * @default vae + * @constant */ - id: string; + type: "vae"; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Base + * @default sdxl + * @constant */ - is_intermediate?: boolean; + base: "sdxl"; + }; + /** ValidationError */ + ValidationError: { + /** Location */ + loc: (string | number)[]; + /** Message */ + msg: string; + /** Error Type */ + type: string; + }; + /** Workflow */ + Workflow: { /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Name + * @description The name of the workflow. */ - use_cache?: boolean; + name: string; /** - * Xyimage Item - * @description The XYImage collection item - * @default null + * Author + * @description The author of the workflow. */ - xyimage_item?: string | null; + author: string; /** - * type - * @default xy_image_expand - * @constant + * Description + * @description The description of the workflow. */ - type: "xy_image_expand"; - }; - /** - * XYImageExpandOutput - * @description XY Image Expand Output - */ - XYImageExpandOutput: { + description: string; + /** + * Version + * @description The version of the workflow. + */ + version: string; /** - * X Item - * @description The X item + * Contact + * @description The contact of the workflow. */ - x_item: string; + contact: string; /** - * Y Item - * @description The y item + * Tags + * @description The tags of the workflow. */ - y_item: string; - /** @description The Image item */ - image: components["schemas"]["ImageField"]; + tags: string; /** - * Width - * @description The width of the image in pixels + * Notes + * @description The notes of the workflow. */ - width: number; + notes: string; /** - * Height - * @description The height of the image in pixels + * Exposedfields + * @description The exposed fields of the workflow. */ - height: number; + exposedFields: components["schemas"]["ExposedField"][]; + /** @description The meta of the workflow. */ + meta: components["schemas"]["WorkflowMeta"]; /** - * type - * @default xy_image_expand_output - * @constant + * Nodes + * @description The nodes of the workflow. */ - type: "xy_image_expand_output"; - }; - /** - * XYImage Tiles To Image - * @description Takes a collection of XYImage Tiles (json of array(x_pos,y_pos,image_name)) and create an image from overlapping tiles - */ - XYImageTilesToImageInvocation: { + nodes: { + [key: string]: components["schemas"]["JsonValue"]; + }[]; /** - * @description The board to save the image to - * @default null + * Edges + * @description The edges of the workflow. */ - board?: components["schemas"]["BoardField"] | null; + edges: { + [key: string]: components["schemas"]["JsonValue"]; + }[]; /** - * @description Optional metadata to be saved with the image - * @default null + * Form + * @description The form of the workflow. */ - metadata?: components["schemas"]["MetadataField"] | null; + form?: { + [key: string]: components["schemas"]["JsonValue"]; + } | null; /** * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * @description The id of the workflow. */ id: string; + }; + /** WorkflowAndGraphResponse */ + WorkflowAndGraphResponse: { /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Workflow + * @description The workflow used to generate the image, as stringified JSON */ - is_intermediate?: boolean; + workflow: string | null; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Graph + * @description The graph used to generate the image, as stringified JSON */ - use_cache?: boolean; + graph: string | null; + }; + /** + * WorkflowCategory + * @enum {string} + */ + WorkflowCategory: "user" | "default"; + /** WorkflowMeta */ + WorkflowMeta: { /** - * Xyimages - * @description The xyImage Collection - * @default [] + * Version + * @description The version of the workflow schema. */ - xyimages?: string[]; + version: string; + /** @description The category of the workflow (user or default). */ + category: components["schemas"]["WorkflowCategory"]; + }; + /** WorkflowRecordDTO */ + WorkflowRecordDTO: { /** - * Blend Mode - * @description Seam blending type Linear or Smart - * @default seam-grad - * @enum {string} + * Workflow Id + * @description The id of the workflow. */ - blend_mode?: "Linear" | "seam-grad" | "seam-sobel1" | "seam-sobel3" | "seam-sobel5" | "seam-sobel7" | "seam-scharr"; + workflow_id: string; /** - * Blur Size - * @description Size of the blur & Gutter to use with Smart Seam - * @default 16 + * Name + * @description The name of the workflow. + */ + name: string; + /** + * Created At + * @description The created timestamp of the workflow. */ - blur_size?: number; + created_at: string; /** - * Search Size - * @description Seam search size in pixels 1-4 are sensible sizes - * @default 1 + * Updated At + * @description The updated timestamp of the workflow. */ - search_size?: number; + updated_at: string; /** - * type - * @default xy_image_tiles_to_image - * @constant + * Opened At + * @description The opened timestamp of the workflow. */ - type: "xy_image_tiles_to_image"; + opened_at?: string | null; + /** @description The workflow. */ + workflow: components["schemas"]["Workflow"]; }; - /** - * XYImages To Grid - * @description Takes Collection of XYImages (json of (x_item,y_item,image_name)array), sorts the images into X,Y and creates a grid image with labels - */ - XYImagesToGridInvocation: { + /** WorkflowRecordListItemWithThumbnailDTO */ + WorkflowRecordListItemWithThumbnailDTO: { /** - * @description The board to save the image to - * @default null + * Workflow Id + * @description The id of the workflow. */ - board?: components["schemas"]["BoardField"] | null; + workflow_id: string; /** - * @description Optional metadata to be saved with the image - * @default null + * Name + * @description The name of the workflow. */ - metadata?: components["schemas"]["MetadataField"] | null; + name: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Created At + * @description The created timestamp of the workflow. */ - id: string; + created_at: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Updated At + * @description The updated timestamp of the workflow. */ - is_intermediate?: boolean; + updated_at: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Opened At + * @description The opened timestamp of the workflow. */ - use_cache?: boolean; + opened_at?: string | null; /** - * Xyimages - * @description The XYImage item Collection - * @default [] + * Description + * @description The description of the workflow. */ - xyimages?: string[]; + description: string; + /** @description The description of the workflow. */ + category: components["schemas"]["WorkflowCategory"]; /** - * Scale Factor - * @description The factor by which to scale the images - * @default 1 + * Tags + * @description The tags of the workflow. */ - scale_factor?: number; + tags: string; /** - * Resample Mode - * @description The resampling mode - * @default bicubic - * @enum {string} + * Thumbnail Url + * @description The URL of the workflow thumbnail. */ - resample_mode?: "nearest" | "box" | "bilinear" | "hamming" | "bicubic" | "lanczos"; + thumbnail_url?: string | null; + }; + /** + * WorkflowRecordOrderBy + * @description The order by options for workflow records + * @enum {string} + */ + WorkflowRecordOrderBy: "created_at" | "updated_at" | "opened_at" | "name"; + /** WorkflowRecordWithThumbnailDTO */ + WorkflowRecordWithThumbnailDTO: { /** - * Left Label Width - * @description Width of the left label area - * @default 100 + * Workflow Id + * @description The id of the workflow. */ - left_label_width?: number; + workflow_id: string; /** - * Label Font Size - * @description Size of the font to use for labels - * @default 16 + * Name + * @description The name of the workflow. */ - label_font_size?: number; + name: string; /** - * type - * @default xy_images_to_grid - * @constant + * Created At + * @description The created timestamp of the workflow. */ - type: "xy_images_to_grid"; - }; - /** - * XY Product CSV - * @description Converts X and Y CSV strings to an XY Item collection with every combination of X and Y - */ - XYProductCSVInvocation: { + created_at: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Updated At + * @description The updated timestamp of the workflow. */ - id: string; + updated_at: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Opened At + * @description The opened timestamp of the workflow. */ - is_intermediate?: boolean; + opened_at?: string | null; + /** @description The workflow. */ + workflow: components["schemas"]["Workflow"]; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Thumbnail Url + * @description The URL of the workflow thumbnail. */ - use_cache?: boolean; + thumbnail_url?: string | null; + }; + /** WorkflowWithoutID */ + WorkflowWithoutID: { /** - * X - * @description x string - * @default null + * Name + * @description The name of the workflow. */ - x?: string | null; + name: string; /** - * Y - * @description y string - * @default null + * Author + * @description The author of the workflow. */ - y?: string | null; + author: string; /** - * type - * @default xy_product_csv - * @constant + * Description + * @description The description of the workflow. */ - type: "xy_product_csv"; - }; - /** - * XY Product - * @description Takes X and Y string collections and outputs a XY Item collection with every combination of X and Y - */ - XYProductInvocation: { + description: string; /** - * Id - * @description The id of this instance of an invocation. Must be unique among all instances of invocations. + * Version + * @description The version of the workflow. */ - id: string; + version: string; /** - * Is Intermediate - * @description Whether or not this is an intermediate invocation. - * @default false + * Contact + * @description The contact of the workflow. */ - is_intermediate?: boolean; + contact: string; /** - * Use Cache - * @description Whether or not to use the cache - * @default true + * Tags + * @description The tags of the workflow. */ - use_cache?: boolean; + tags: string; /** - * X Collection - * @description The X collection - * @default [] + * Notes + * @description The notes of the workflow. */ - x_collection?: string[]; + notes: string; /** - * Y Collection - * @description The Y collection - * @default [] + * Exposedfields + * @description The exposed fields of the workflow. */ - y_collection?: string[]; + exposedFields: components["schemas"]["ExposedField"][]; + /** @description The meta of the workflow. */ + meta: components["schemas"]["WorkflowMeta"]; /** - * type - * @default xy_product - * @constant + * Nodes + * @description The nodes of the workflow. */ - type: "xy_product"; - }; - /** - * XYProductOutput - * @description XYCProductOutput a collection that contains every combination of the input collections - */ - XYProductOutput: { + nodes: { + [key: string]: components["schemas"]["JsonValue"]; + }[]; /** - * Xy Item Collection - * @description The XY Item collection + * Edges + * @description The edges of the workflow. */ - xy_item_collection: string[]; + edges: { + [key: string]: components["schemas"]["JsonValue"]; + }[]; /** - * type - * @default xy_collect_output - * @constant + * Form + * @description The form of the workflow. */ - type: "xy_collect_output"; + form?: { + [key: string]: components["schemas"]["JsonValue"]; + } | null; }; /** * ZImageConditioningField From 6691871d5fc2f0bf83abcd5f3fe409031dadffbc Mon Sep 17 00:00:00 2001 From: skunkworxdark Date: Sun, 5 Apr 2026 21:33:26 +0100 Subject: [PATCH 6/8] another typegen fix --- invokeai/frontend/web/src/services/api/schema.ts | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/invokeai/frontend/web/src/services/api/schema.ts b/invokeai/frontend/web/src/services/api/schema.ts index 93c0159ed07..bc72a5e8073 100644 --- a/invokeai/frontend/web/src/services/api/schema.ts +++ b/invokeai/frontend/web/src/services/api/schema.ts @@ -14575,14 +14575,14 @@ export type components = { * Convert Cache Dir * Format: path * @description Path to the converted models cache directory (DEPRECATED, but do not delete because it is needed for migration from previous versions). - * @default models\.convert_cache + * @default models/.convert_cache */ convert_cache_dir?: string; /** * Download Cache Dir * Format: path * @description Path to the directory that contains dynamically downloaded models. - * @default models\.download_cache + * @default models/.download_cache */ download_cache_dir?: string; /** From 326a88c2c76d1325750e47effb311d858402222e Mon Sep 17 00:00:00 2001 From: skunkworxdark Date: Thu, 9 Apr 2026 17:41:43 +0100 Subject: [PATCH 7/8] refactor(ui): consolidate model filter and sort controls into a unified menu - Replaced separate `ModelSortControl` and `ModelTypeFilter` components with a single, unified "Filtering" dropdown menu. - Organised filtering options into categorised submenus in the following order: Direction, Sort By, and Model Type. - Enhanced submenu labels to display the currently active selection inline for quick reference. - Improved visual alignment within menus by using hidden checkmarks on unselected items, ensuring consistent indentation across all options. - Resolved styling and linting issues (unused variables, JSX bind warnings) within the new component. --- .../web/src/common/hooks/useSubMenu.tsx | 11 +- .../ModelManagerPanel/ModelFilterMenu.tsx | 238 ++++++++++++++++++ .../ModelManagerPanel/ModelListNavigation.tsx | 8 +- .../ModelManagerPanel/ModelSortControl.tsx | 85 ------- .../ModelManagerPanel/ModelTypeFilter.tsx | 78 ------ 5 files changed, 249 insertions(+), 171 deletions(-) create mode 100644 invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelFilterMenu.tsx delete mode 100644 invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx delete mode 100644 invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelTypeFilter.tsx diff --git a/invokeai/frontend/web/src/common/hooks/useSubMenu.tsx b/invokeai/frontend/web/src/common/hooks/useSubMenu.tsx index f8ea01909a7..4c1bc56e495 100644 --- a/invokeai/frontend/web/src/common/hooks/useSubMenu.tsx +++ b/invokeai/frontend/web/src/common/hooks/useSubMenu.tsx @@ -151,11 +151,18 @@ export const useSubMenu = (): UseSubMenuReturn => { }; }; -export const SubMenuButtonContent = ({ label }: { label: string }) => { +export const SubMenuButtonContent = ({ label, value }: { label: string; value?: string }) => { return ( {label} - + + {value !== undefined && ( + + {value} + + )} + + ); }; diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelFilterMenu.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelFilterMenu.tsx new file mode 100644 index 00000000000..370ffc05415 --- /dev/null +++ b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelFilterMenu.tsx @@ -0,0 +1,238 @@ +import { Button, Flex, Menu, MenuButton, MenuItem, MenuList } from '@invoke-ai/ui-library'; +import { useAppDispatch, useAppSelector } from 'app/store/storeHooks'; +import { SubMenuButtonContent, useSubMenu } from 'common/hooks/useSubMenu'; +import type { ModelCategoryData } from 'features/modelManagerV2/models'; +import { MODEL_CATEGORIES_AS_LIST } from 'features/modelManagerV2/models'; +import { + selectFilteredModelType, + selectOrderBy, + selectSortDirection, + setFilteredModelType, + setOrderBy, + setSortDirection, +} from 'features/modelManagerV2/store/modelManagerV2Slice'; +import { memo, useCallback, useMemo } from 'react'; +import { useTranslation } from 'react-i18next'; +import { + PiCheckBold, + PiFunnelBold, + PiListBold, + PiSortAscendingBold, + PiSortDescendingBold, + PiWarningBold, +} from 'react-icons/pi'; + +type OrderBy = 'default' | 'name' | 'type' | 'base' | 'size' | 'created_at' | 'updated_at' | 'path' | 'format'; + +const ORDER_BY_OPTIONS: OrderBy[] = [ + 'default', + 'name', + 'base', + 'size', + 'created_at', + 'updated_at', + 'path', + 'type', + 'format', +]; + +const SortByMenuItem = memo(({ option, label }: { option: OrderBy; label: string }) => { + const dispatch = useAppDispatch(); + const orderBy = useAppSelector(selectOrderBy); + const onClick = useCallback(() => { + dispatch(setOrderBy(option)); + }, [dispatch, option]); + + return ( + : } + > + {label} + + ); +}); +SortByMenuItem.displayName = 'SortByMenuItem'; + +const SortBySubMenu = memo(() => { + const { t } = useTranslation(); + const subMenu = useSubMenu(); + const orderBy = useAppSelector(selectOrderBy); + + const ORDER_BY_LABELS = useMemo( + () => ({ + default: t('modelManager.sortDefault'), + name: t('modelManager.sortByName'), + base: t('modelManager.sortByBase'), + size: t('modelManager.sortBySize'), + created_at: t('modelManager.sortByDateAdded'), + updated_at: t('modelManager.sortByDateModified'), + path: t('modelManager.sortByPath'), + type: t('modelManager.sortByType'), + format: t('modelManager.sortByFormat'), + }), + [t] + ); + + return ( + }> + + + + + + {ORDER_BY_OPTIONS.map((option) => ( + + ))} + + + + ); +}); +SortBySubMenu.displayName = 'SortBySubMenu'; + +const DirectionSubMenu = memo(() => { + const { t } = useTranslation(); + const dispatch = useAppDispatch(); + const direction = useAppSelector(selectSortDirection); + const subMenu = useSubMenu(); + + const setDirectionAsc = useCallback(() => { + dispatch(setSortDirection('asc')); + }, [dispatch]); + + const setDirectionDesc = useCallback(() => { + dispatch(setSortDirection('desc')); + }, [dispatch]); + + const currentValue = direction === 'asc' ? t('common.ascending', 'Ascending') : t('common.descending', 'Descending'); + + return ( + : } + > + + + + + + : } + > + {t('common.ascending', 'Ascending')} + + : } + > + {t('common.descending', 'Descending')} + + + + + ); +}); +DirectionSubMenu.displayName = 'DirectionSubMenu'; + +const ModelTypeSubMenu = memo(() => { + const { t } = useTranslation(); + const dispatch = useAppDispatch(); + const filteredModelType = useAppSelector(selectFilteredModelType); + const subMenu = useSubMenu(); + + const clearModelType = useCallback(() => { + dispatch(setFilteredModelType(null)); + }, [dispatch]); + + const setMissingFilter = useCallback(() => { + dispatch(setFilteredModelType('missing')); + }, [dispatch]); + + const currentValue = useMemo(() => { + if (filteredModelType === null) { + return t('modelManager.allModels'); + } + if (filteredModelType === 'missing') { + return t('modelManager.missingFiles'); + } + const categoryData = MODEL_CATEGORIES_AS_LIST.find((data) => data.category === filteredModelType); + return categoryData ? t(categoryData.i18nKey) : ''; + }, [filteredModelType, t]); + + return ( + }> + + + + + + : } + > + {t('modelManager.allModels')} + + : } + > + + {filteredModelType !== 'missing' && } + {t('modelManager.missingFiles')} + + + {MODEL_CATEGORIES_AS_LIST.map((data) => ( + + ))} + + + + ); +}); +ModelTypeSubMenu.displayName = 'ModelTypeSubMenu'; + +const ModelMenuItem = memo(({ data }: { data: ModelCategoryData }) => { + const { t } = useTranslation(); + const dispatch = useAppDispatch(); + const filteredModelType = useAppSelector(selectFilteredModelType); + const onClick = useCallback(() => { + dispatch(setFilteredModelType(data.category)); + }, [data.category, dispatch]); + return ( + : } + > + {t(data.i18nKey)} + + ); +}); +ModelMenuItem.displayName = 'ModelMenuItem'; + +export const ModelFilterMenu = memo(() => { + const { t } = useTranslation(); + + return ( + + }> + {t('common.filtering', 'Filtering')} + + + + + + + + ); +}); + +ModelFilterMenu.displayName = 'ModelFilterMenu'; diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx index 2f3dcb94efe..bbfb88df5cf 100644 --- a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx +++ b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelListNavigation.tsx @@ -6,9 +6,8 @@ import type { ChangeEventHandler } from 'react'; import { memo, useCallback } from 'react'; import { PiXBold } from 'react-icons/pi'; +import { ModelFilterMenu } from './ModelFilterMenu'; import { ModelListBulkActions } from './ModelListBulkActions'; -import { ModelSortControl } from './ModelSortControl'; -import { ModelTypeFilter } from './ModelTypeFilter'; export const ModelListNavigation = memo(() => { const dispatch = useAppDispatch(); @@ -51,10 +50,7 @@ export const ModelListNavigation = memo(() => { - - - - + diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx deleted file mode 100644 index 1e84de30dbf..00000000000 --- a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelSortControl.tsx +++ /dev/null @@ -1,85 +0,0 @@ -import { Flex, IconButton, Select } from '@invoke-ai/ui-library'; -import { useAppDispatch, useAppSelector } from 'app/store/storeHooks'; -import { - selectOrderBy, - selectSortDirection, - setOrderBy, - setSortDirection, -} from 'features/modelManagerV2/store/modelManagerV2Slice'; -import type { ChangeEvent } from 'react'; -import { useCallback, useMemo } from 'react'; -import { useTranslation } from 'react-i18next'; -import { PiSortAscendingBold, PiSortDescendingBold } from 'react-icons/pi'; -import { z } from 'zod'; - -const zOrderBy = z.enum(['default', 'name', 'type', 'base', 'size', 'created_at', 'updated_at', 'path', 'format']); -type OrderBy = z.infer; -const isOrderBy = (v: unknown): v is OrderBy => zOrderBy.safeParse(v).success; - -const ORDER_BY_OPTIONS: OrderBy[] = [ - 'default', - 'name', - 'base', - 'size', - 'created_at', - 'updated_at', - 'path', - 'type', - 'format', -]; - -export const ModelSortControl = () => { - const { t } = useTranslation(); - const dispatch = useAppDispatch(); - - const orderBy = useAppSelector(selectOrderBy); - const direction = useAppSelector(selectSortDirection); - - const ORDER_BY_LABELS = useMemo( - () => ({ - default: t('modelManager.sortDefault'), - name: t('modelManager.sortByName'), - base: t('modelManager.sortByBase'), - size: t('modelManager.sortBySize'), - created_at: t('modelManager.sortByDateAdded'), - updated_at: t('modelManager.sortByDateModified'), - path: t('modelManager.sortByPath'), - type: t('modelManager.sortByType'), - format: t('modelManager.sortByFormat'), - }), - [t] - ); - - const onChangeOrderBy = useCallback( - (e: ChangeEvent) => { - if (!isOrderBy(e.target.value)) { - return; - } - dispatch(setOrderBy(e.target.value)); - }, - [dispatch] - ); - - const toggleDirection = useCallback(() => { - dispatch(setSortDirection(direction === 'asc' ? 'desc' : 'asc')); - }, [dispatch, direction]); - - return ( - - - : } - size="sm" - variant="ghost" - onClick={toggleDirection} - /> - - ); -}; diff --git a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelTypeFilter.tsx b/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelTypeFilter.tsx deleted file mode 100644 index 5aa8e628869..00000000000 --- a/invokeai/frontend/web/src/features/modelManagerV2/subpanels/ModelManagerPanel/ModelTypeFilter.tsx +++ /dev/null @@ -1,78 +0,0 @@ -import { Button, Flex, Menu, MenuButton, MenuItem, MenuList } from '@invoke-ai/ui-library'; -import { useAppDispatch, useAppSelector } from 'app/store/storeHooks'; -import type { ModelCategoryData } from 'features/modelManagerV2/models'; -import { MODEL_CATEGORIES, MODEL_CATEGORIES_AS_LIST } from 'features/modelManagerV2/models'; -import type { ModelCategoryType } from 'features/modelManagerV2/store/modelManagerV2Slice'; -import { selectFilteredModelType, setFilteredModelType } from 'features/modelManagerV2/store/modelManagerV2Slice'; -import { memo, useCallback } from 'react'; -import { useTranslation } from 'react-i18next'; -import { PiFunnelBold, PiWarningBold } from 'react-icons/pi'; - -const isModelCategoryType = (type: string): type is ModelCategoryType => { - return type in MODEL_CATEGORIES; -}; - -export const ModelTypeFilter = memo(() => { - const { t } = useTranslation(); - const dispatch = useAppDispatch(); - const filteredModelType = useAppSelector(selectFilteredModelType); - - const clearModelType = useCallback(() => { - dispatch(setFilteredModelType(null)); - }, [dispatch]); - - const setMissingFilter = useCallback(() => { - dispatch(setFilteredModelType('missing')); - }, [dispatch]); - - const getButtonLabel = () => { - if (filteredModelType === 'missing') { - return t('modelManager.missingFiles'); - } - if (filteredModelType && isModelCategoryType(filteredModelType)) { - return t(MODEL_CATEGORIES[filteredModelType].i18nKey); - } - return t('modelManager.allModels'); - }; - - return ( - - }> - {getButtonLabel()} - - - {t('modelManager.allModels')} - - - - {t('modelManager.missingFiles')} - - - {MODEL_CATEGORIES_AS_LIST.map((data) => ( - - ))} - - - ); -}); - -ModelTypeFilter.displayName = 'ModelTypeFilter'; - -const ModelMenuItem = memo(({ data }: { data: ModelCategoryData }) => { - const { t } = useTranslation(); - const dispatch = useAppDispatch(); - const filteredModelType = useAppSelector(selectFilteredModelType); - const onClick = useCallback(() => { - dispatch(setFilteredModelType(data.category)); - }, [data.category, dispatch]); - return ( - - {t(data.i18nKey)} - - ); -}); -ModelMenuItem.displayName = 'ModelMenuItem'; From 262a4e2d1eecb3753679aa869223d42e40bf5a29 Mon Sep 17 00:00:00 2001 From: skunkworxdark Date: Thu, 9 Apr 2026 18:01:02 +0100 Subject: [PATCH 8/8] Lint fix --- invokeai/frontend/web/src/features/modelManagerV2/models.ts | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/invokeai/frontend/web/src/features/modelManagerV2/models.ts b/invokeai/frontend/web/src/features/modelManagerV2/models.ts index 7b5a08adfe2..ea8fe69481d 100644 --- a/invokeai/frontend/web/src/features/modelManagerV2/models.ts +++ b/invokeai/frontend/web/src/features/modelManagerV2/models.ts @@ -30,7 +30,7 @@ export type ModelCategoryData = { filter: (config: AnyModelConfig) => boolean; }; -export const MODEL_CATEGORIES: Record = { +const MODEL_CATEGORIES: Record = { unknown: { category: 'unknown', i18nKey: 'common.unknown',