A powerful toolkit for implementing AI generation providers in CreativeEditor SDK.
Note: This package is only relevant if you need to create new AI providers or extend existing functionality. For simple integration of AI features, use the @imgly/plugin-ai-apps-web package instead.
This package provides the foundation for creating AI generation plugins for CreativeEditor SDK. It offers a standardized interface for implementing AI generation providers that can create images, videos, audio, or text assets. The package includes utilities for handling:
- Provider registration and initialization
- User interface generation
- Global action registry for quick actions and plugin actions
- Type-safe quick action definitions
- Cross-plugin action support
npm install @imgly/plugin-ai-generation-web
The core of this package is the Provider
interface which defines the contract for AI generation providers. Here's how to implement a basic provider:
import {
Provider,
ImageOutput,
initializeProvider,
loggingMiddleware,
CommonProviderConfiguration
} from '@imgly/plugin-ai-generation-web';
// Define your provider configuration interface
interface MyProviderConfiguration
extends CommonProviderConfiguration<MyInputType, ImageOutput> {
// Add any provider-specific configuration here
baseURL?: string;
}
// Create a provider factory function
function createMyImageProvider(config: MyProviderConfiguration): Provider<'image', MyInputType, ImageOutput> {
return {
// Unique identifier for this provider
id: 'my-image-provider',
// Define output asset type, other options are 'video', 'audio', 'text'
kind: 'image',
// Initialize the provider
initialize: async ({ engine, cesdk }) => {
// Setup APIs, register further components, etc.
myAIApi.configure({
apiKey: 'YOUR_API_KEY',
headers: config.headers // Use custom headers if provided
});
},
// Define input panel and UI components
input: {
// Define how the input panel is rendered
panel: {
// Option 1: Schema-based UI (using OpenAPI)
type: 'schema',
document: myApiSchema,
inputReference: '#/components/schemas/GenerationInput',
getBlockInput: async (input) => ({
image: { width: 1024, height: 1024 }
})
},
// Quick actions supported by this provider
quickActions: {
supported: {
'ly.img.editImage': {
mapInput: (input) => ({
prompt: input.prompt,
image_url: input.uri
})
},
'ly.img.styleTransfer': {
mapInput: (input) => ({
prompt: input.style,
image_url: input.uri
})
}
}
}
},
// Define output generation behavior
output: {
// Allow cancellation
abortable: true,
// Store generated assets, options are:
// - false: No history
// - '@imgly/local': In-memory storage (lost on refresh)
// - '@imgly/indexedDB': Browser IndexedDB storage
// - any other string: Handled as a custom asset source ID
history: '@imgly/indexedDB',
// Add middleware for pre/post-processing of the generation
middleware: [loggingMiddleware()],
// Configure success/error notifications
notification: {
success: {
show: true,
message: 'Generation successful!'
}
},
// Core generation function
generate: async (input, { abortSignal, engine }) => {
// Call your AI API and return result
const response = await myAIApi.generateImage(input, {
headers: config.headers // Pass custom headers to API
});
return {
kind: 'image',
url: response.imageUrl
};
}
}
};
}
// Usage example
const myImageProvider = createMyImageProvider({
proxyUrl: 'http://your-proxy-server.com/api/proxy',
headers: {
'x-client-version': '1.0.0',
'x-request-source': 'cesdk-plugin'
},
debug: false,
middleware: [loggingMiddleware()],
baseURL: 'https://assets.example.com'
});
The package includes a global ActionRegistry
for managing quick actions and plugin actions. To register a new action:
import { ActionRegistry } from '@imgly/plugin-ai-generation-web';
// Get the global registry instance
const registry = ActionRegistry.get();
// Register a quick action
const unregister = registry.register({
id: 'my-quick-action',
type: 'quick',
kind: 'image',
label: 'My Quick Action',
enable: true,
render: (context) => {
// Render the quick action UI
context.builder.Button('my-button', {
label: 'Generate',
onClick: async () => {
await context.generate({ prompt: 'Hello world' });
}
});
}
});
The Provider interface is generic and type-safe, supporting four output kinds:
// K: Output kind ('image', 'video', 'audio', 'text')
// I: Input type specific to your provider, i.e. what does the generate function need
// O: Output type (ImageOutput, VideoOutput, AudioOutput, TextOutput)
// C: Chunk type for streaming (optional, defaults to O)
interface Provider<K extends OutputKind, I, O extends Output, C = O> { ... }
All providers should extend the CommonProviderConfiguration
interface, which provides standardized configuration options:
interface CommonProviderConfiguration<I, O extends Output> {
// The proxy URL to use for the provider
proxyUrl: string;
// Enable debug mode for additional logging
debug?: boolean;
// Middleware for request/response processing
middleware?: Middleware<I, O>[];
// Custom headers to include in all API requests
headers?: Record<string, string>;
}
The headers
property allows you to include custom HTTP headers in all API requests made by your provider. This is useful for:
- Adding custom client identification headers
- Including version information
- Passing through metadata required by your API
- Adding correlation IDs for request tracing
Implementation Note: When implementing your provider's generate
function, ensure you merge the custom headers with any required headers for your API:
// In your generate function
const response = await fetch(apiUrl, {
method: 'POST',
headers: {
'Content-Type': 'application/json',
'Authorization': `Bearer ${apiKey}`,
...config.headers // Spread custom headers
},
body: JSON.stringify(requestData)
});
- id: Unique identifier for your provider
- kind: Type of asset generated ('image', 'video', 'audio', 'text')
- name: Optional human-readable name
- initialize: Setup function called when the provider is loaded
- input: Configuration for input UI and parameters
- output: Configuration for generation and result handling
The output
property has several important options:
- generate: Main function that performs the actual generation
- history: Asset storage strategy ('false', '@imgly/local', '@imgly/indexedDB', or custom ID)
- abortable: Whether generation can be cancelled by the user
- middleware: Array of middleware functions for pre/post-processing
- notification: Success and error notification configuration
- generationHintText: Text to display below the generation button
The notification system allows fine-grained control over success and error messages:
notification: {
success: {
// Control whether to show notifications (can be dynamic)
show: true, // or (context) => shouldShow(context)
// Message text or i18n key (can be dynamic)
message: 'Generation successful!', // or (context) => getMessage(context)
// Optional action button
action: {
label: 'View', // or (context) => getLabel(context)
onClick: (context) => { /* handle click */ }
},
// How long to show the notification
duration: 'short' // or 'medium', 'long', 'infinite'
},
error: {
// Similar options for error notifications
show: true,
message: 'Generation failed', // or (context) => getErrorMessage(context)
// ...
}
}
The generate
function can return a simple output object or an AsyncGenerator for streaming results:
// Simple response
generate: async (input, options) => {
const result = await api.generateImage(input);
return { kind: 'image', url: result.url };
}
// Streaming response (currently only supported for text)
generate: async function* (input, options) {
const stream = api.streamGenerationResult(input);
let inferredText: string = '';
// Yield interim results
for await (const chunk of stream) {
inferredText += chunk;
yield { kind: 'text', text: inferredText };
}
// Return final result
return { kind: 'text', text: inferredText };
}
The generationHintText
property allows providers to display helpful information below the generation button:
generationHintText: "Generation may take up to a minute. You can close this panel and will be notified when ready."
The package supports two approaches for creating input panels:
The schema
type uses OpenAPI specification to declaratively define your input form.
input: {
panel: {
type: 'schema',
// Complete OpenAPI v3 document describing your inputs
document: myOpenAPISchema,
// JSON pointer to your input schema within the document
inputReference: '#/components/schemas/GenerationInput',
// Optional property to control display order
orderExtensionKeyword: 'x-order-properties',
// Function that converts input to block parameters
getBlockInput: async (input) => ({
image: { width: 1024, height: 1024 }
}),
// Optional custom renderers for specific properties found in the schema
renderCustomProperty: {
// This is a custom renderer for a fictional `imageUrl` property
imageUrl: (context, property) => {
const valueState = context.state('imageUrl', '');
context.builder.TextInput('imageUrl', {
inputLabel: 'Image URL',
...valueState
});
// Return a function that returns the value for this property
return () => { id: property.id, type: 'string', value: valueState.value };
}
}
}
}
{
"openapi": "3.0.0",
"components": {
"schemas": {
"GenerationInput": {
"type": "object",
"required": ["prompt"],
"properties": {
"prompt": {
"type": "string",
"title": "Prompt",
"description": "Describe what you want to generate",
"x-imgly-builder": {
"component": "TextArea"
}
},
"width": {
"type": "integer",
"title": "Width",
"default": 1024,
"enum": [512, 1024, 2048],
"x-imgly-builder": {
"component": "Select"
}
}
},
"x-order-properties": ["prompt", "width"]
}
}
}
}
- Built-in validation based on schema constraints
- AI providers like fal.ai provide schemas for their models
- Automatic UI component generation based on property types
- Extensions like
x-imgly-builder
to specify component types - Property ordering via
orderExtensionKeyword
- Customizable property rendering with
renderCustomProperty
The custom
type gives you complete control over UI components. For more details on how to build custom panels and see all available builder components, refer to the Create a Custom Panel guide.
input: {
panel: {
type: 'custom',
render: (context, options) => {
// Use the builder pattern to create UI components
const promptState = context.state('prompt', '');
context.builder.TextArea('prompt', {
inputLabel: 'Prompt',
...promptState
});
// Set up width selection
const widthState = context.state('width', 1024);
context.builder.Select('width', {
inputLabel: 'Width',
options: [
{ value: 512, label: '512px' },
{ value: 1024, label: '1024px' },
{ value: 2048, label: '2048px' }
],
...widthState
});
// Return functions to get input values and block parameters
return {
// The input for the generate function
getInput: () => ({
prompt: promptState.value,
width: widthState.value
}),
// The input for the block creation
getBlockInput: () => ({
image: {
width: widthState.value,
height: widthState.value,
label: `AI Image: ${promptState.value.substring(0, 20)}...`
}
})
};
}
}
}
- Complete control over UI components and layout
- Complex logic between fields (dependencies, conditionals)
- Dynamic UI that changes based on user interactions
Both panel types accept additional configuration:
panel: {
type: 'schema', // or 'custom'
// ...panel type specific options
// Control the generation flow
userFlow: 'placeholder', // or 'generation-only' (default)
// Include/exclude history library from panel
includeHistoryLibrary: true // (default)
}
-
userFlow:
-
placeholder
: Creates a block as a placeholder with loading state when generation starts -
generation-only
: Only triggers generation without creating a placeholder
-
-
includeHistoryLibrary: Controls whether the history library is shown in the panel
The getBlockInput
function is crucial for both panel types. It converts your input into the parameters needed to create a block in CreativeEditor SDK.
- Defines dimensions, duration, and appearance of asset blocks
- Creates placeholders before generation completes
- Maps your AI provider's inputs to standardized block parameters
Each output kind requires specific parameters:
getBlockInput: async (input) => ({
image: {
width: 1024, // Required - Width in pixels
height: 1024, // Required - Height in pixels
label: 'My Image' // Optional - Display name
}
});
getBlockInput: async (input) => ({
video: {
width: 1280, // Required - Width in pixels
height: 720, // Required - Height in pixels
duration: 10, // Required - Duration in seconds
label: 'My Video' // Optional - Display name
}
});
getBlockInput: async (input) => ({
audio: {
duration: 30, // Optional - Duration in seconds
thumbnailUrl: 'path/to/img.jpg', // Optional - URL for thumbnail
label: 'My Audio' // Optional - Display name
}
});
getBlockInput: async (input) => ({
text: {
length: 250, // Required - Approximate character length
label: 'My Text' // Optional - Display name
}
});
Quick Actions provide context-aware AI generation capabilities directly in CreativeEditor SDK's canvas menu. Unlike panels (which appear in the side panel), quick actions appear when users select elements on the canvas.
Here are all the quick action IDs that can be used in the supported
field of your provider configuration:
-
ly.img.artistTransfer
: Transform image in the style of famous artists- Input:
{ artist: string, uri: string }
- Input:
-
ly.img.combineImages
: Combine multiple images with instructions- Input:
{ prompt: string, uris: string[], exportFromBlockIds: number[] }
- Input:
-
ly.img.createVariant
: Create a variation of the image- Input:
{ prompt: string, uri: string }
- Input:
-
ly.img.editImage
: Change image based on description- Input:
{ prompt: string, uri: string }
- Input:
-
ly.img.remixPage
: Convert the page into a single image- Input:
{ prompt: string, uri: string }
- Input:
-
ly.img.remixPageWithPrompt
: Remix the page with custom instructions- Input:
{ prompt: string, uri: string }
- Input:
-
ly.img.styleTransfer
: Transform image into different art styles- Input:
{ style: string, uri: string }
- Input:
-
ly.img.swapBackground
: Change the background of the image- Input:
{ prompt: string, uri: string }
- Input:
-
ly.img.gpt-image-1.changeStyleLibrary
: Apply different art styles (GPT-specific)- Input:
{ prompt: string, uri: string }
- Input:
-
ly.img.changeTextTo
: Change text to a different format or style- Input:
{ prompt: string, customPrompt: string }
- Input:
-
ly.img.changeTone
: Change the tone of the text- Input:
{ prompt: string, type: string }
- Input:
-
ly.img.fix
: Fix spelling and grammar- Input:
{ prompt: string }
- Input:
-
ly.img.improve
: Improve writing quality- Input:
{ prompt: string }
- Input:
-
ly.img.longer
: Make text longer- Input:
{ prompt: string }
- Input:
-
ly.img.shorter
: Make text shorter- Input:
{ prompt: string }
- Input:
-
ly.img.translate
: Translate text to different languages- Input:
{ prompt: string, language: string }
- Input:
-
ly.img.createVideo
: Opens the image2video generation panel with the current image- Input:
{ uri: string }
- Input:
Providers declare which quick actions they support and how to map quick action inputs to provider inputs:
const myProvider = {
// ... other provider config
input: {
// ... panel config
quickActions: {
supported: {
'ly.img.editImage': {
mapInput: (quickActionInput) => ({
prompt: quickActionInput.prompt,
image_url: quickActionInput.uri
})
},
'ly.img.styleTransfer': {
mapInput: (quickActionInput) => ({
style: quickActionInput.style,
image_url: quickActionInput.uri
})
}
}
}
}
};
Quick actions can have two rendering modes:
- Collapsed View: Shows as a simple button in the quick action menu alongside other actions
- Expanded View: Takes over the entire menu space, hiding other actions while the user interacts with this specific action
The expanded view is useful for quick actions that need user input (like text prompts). When a quick action is expanded, the complete menu is replaced with the expanded interface, and other menu items are not shown until the user either completes the action or cancels back to the collapsed view.
render: ({ builder, isExpanded, toggleExpand }) => {
if (isExpanded) {
// Expanded view - takes over the entire menu
builder.TextArea('prompt', { /* input fields */ });
builder.ButtonRow('actions', { /* confirm/cancel buttons */ });
} else {
// Collapsed view - simple button alongside other actions
builder.Button('expand', {
label: 'Edit Image...',
onClick: toggleExpand
});
}
}
Once you've created your provider, you need to initialize it with CreativeEditor SDK and integrate it into the UI.
Use the initializeProvider
function to register your provider:
import { initializeProvider } from '@imgly/plugin-ai-generation-web';
// Create your provider
const myProvider = createMyProvider({
proxyUrl: 'http://your-proxy-server.com/api/proxy',
headers: {
'x-custom-header': 'value',
'x-client-version': '1.0.0'
}
});
// Initialize the provider
function setupMyProvider(cesdk) {
const result = initializeProvider(
myProvider,
{
engine: cesdk.engine,
cesdk
},
{
debug: false,
dryRun: false
}
);
return result;
}
When a provider is initialized, it automatically registers panels with specific IDs:
ly.img.ai.{provider-id}
For example:
- A provider with ID
my-image-provider
registers a panel with IDly.img.ai.my-image-provider
- A provider with ID
fal-ai/recraft-v3
registers a panel with IDly.img.ai.fal-ai/recraft-v3
You can programmatically get a panel ID using the getPanelId
function:
import { getPanelId } from '@imgly/plugin-ai-generation-web';
// Get panel ID for a provider
const panelId = getPanelId('my-image-provider');
// Open the panel
cesdk.ui.openPanel(panelId);
Quick actions are automatically registered in canvas menus with these IDs:
ly.img.ai.{kind}.canvasMenu
For example:
- Image quick actions:
ly.img.ai.image.canvasMenu
- Video quick actions:
ly.img.ai.video.canvasMenu
- Audio quick actions:
ly.img.ai.audio.canvasMenu
- Text quick actions:
ly.img.ai.text.canvasMenu
IMG.LY offers several pre-built AI generation packages that work with this base plugin:
import CreativeEditorSDK from '@cesdk/cesdk-js';
// Import plugin packages
import ImageGeneration from '@imgly/plugin-ai-image-generation-web';
import FalAiImage from '@imgly/plugin-ai-image-generation-web/fal-ai';
import VideoGeneration from '@imgly/plugin-ai-video-generation-web';
import FalAiVideo from '@imgly/plugin-ai-video-generation-web/fal-ai';
// Initialize CreativeEditor SDK
CreativeEditorSDK.create(domElement, {
license: 'your-license-key'
}).then(async (cesdk) => {
// Add default asset sources
await cesdk.addDefaultAssetSources();
// Image generation with Fal.ai models
cesdk.addPlugin(
ImageGeneration({
text2image: FalAiImage.RecraftV3({
proxyUrl: 'http://your-proxy-server.com/api/proxy'
}),
// Alternative: FalAiImage.Recraft20b({ proxyUrl: 'http://your-proxy-server.com/api/proxy' }),
image2image: FalAiImage.GeminiFlashEdit({
proxyUrl: 'http://your-proxy-server.com/api/proxy'
})
})
);
// Video generation
cesdk.addPlugin(
VideoGeneration({
text2video: FalAiVideo.MinimaxVideo01Live({
proxyUrl: 'http://your-proxy-server.com/api/proxy'
})
})
);
// Add quick action menus to canvas
cesdk.ui.setCanvasMenuOrder([
'ly.img.ai.image.canvasMenu',
'ly.img.ai.video.canvasMenu',
...cesdk.ui.getCanvasMenuOrder()
]);
});
The package includes a middleware system to augment the generation flow:
import { rateLimitMiddleware } from '@imgly/plugin-ai-generation-web';
// Create a rate limiting middleware
const rateLimit = rateLimitMiddleware({
maxRequests: 10,
timeWindowMs: 60000, // 1 minute
onRateLimitExceeded: (input, options, info) => {
console.log(
`Rate limit exceeded: ${info.currentCount}/${info.maxRequests}`
);
return false; // Reject request
}
});
// Apply middleware to your provider
const provider = {
// ...provider config
output: {
middleware: [rateLimit]
// ...other output config
}
};
Note: This middleware provides client-side rate limiting for UI purposes only. Always implement proper server-side rate limiting and authentication for production APIs.
The uploadMiddleware
allows you to upload generated content to your own servers:
import { uploadMiddleware } from '@imgly/plugin-ai-generation-web';
// Create an upload middleware
const upload = uploadMiddleware(async (output) => {
// Upload the output to your server/storage
const response = await fetch('https://your-api.example.com/upload', {
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify(output)
});
const result = await response.json();
// Return the output with the updated URL
return {
...output,
url: result.url
};
});
// Apply middleware to your provider
const provider = {
// ...provider config
output: {
middleware: [upload]
// ...other output config
}
};
The ProviderRegistry
is a global singleton that manages all registered providers:
import { ProviderRegistry } from '@imgly/plugin-ai-generation-web';
// Get the global registry
const registry = ProviderRegistry.get();
// Get all registered providers
const allProviders = registry.getAll();
// Get providers by kind
const imageProviders = registry.getByKind('image');
// Find a specific provider
const myProvider = registry.getById('my-provider-id');
This package is fully typed with TypeScript, providing excellent IntelliSense support during development:
- Generic Provider Types: Strongly typed providers with input/output validation
- Quick Action Types: Type-safe quick action definitions with proper input mapping
- Registry Types: Fully typed action and provider registries
- Middleware Types: Typed middleware functions for better composition
// Provider types and interfaces
export { Provider, ImageOutput, VideoOutput, AudioOutput, TextOutput } from './core/provider';
// Action registry
export { ActionRegistry, QuickActionDefinition, PluginActionDefinition } from './core/ActionRegistry';
// Provider registry
export { ProviderRegistry } from './core/ProviderRegistry';
// Initialization functions
export { initializeProvider, initializeProviders } from './providers/';
// Middleware
export { loggingMiddleware, rateLimitMiddleware, uploadMiddleware } from './middleware/';
// Utilities
export { getPanelId, enableQuickActionForImageFill } from './utils/';
// Provider configuration
interface CommonProviderConfiguration<I, O extends Output> {
proxyUrl: string;
debug?: boolean;
middleware?: Middleware<I, O>[];
headers?: Record<string, string>;
}
// Quick action definition
interface QuickActionDefinition<Q extends Record<string, any>> {
id: string;
type: 'quick';
kind: OutputKind;
label?: string;
enable: boolean | ((context: { engine: CreativeEngine }) => boolean);
render: (context: QuickActionRenderContext<Q>) => void;
}