Utility Nodes
Utility nodes often focus on models that can make generated outputs more stylistically consistent with both aesthetic and technical requirements of a production
Utility Nodes
Utility Nodes provide general-purpose functions that support all Atlas workflows. They are not tied to image or mesh generation; instead, they handle documents, text processing, arrays, and lightweight LLM operations.
Utility Nodes are essential for orchestrating complex workflows, preparing inputs, and structuring data for generation nodes.
Document Nodes
Extract Images From Document
This node processes an uploaded PDF document and extracts all images embedded within it.
Input: PDF file (e.g., a Game Design Document)
Output: Image Array
Each extracted image is returned as an element in the array.
You can use the resulting array with:
Find Images by Description node to pull specific images
Image Generation nodes that accept image arrays
Break Images Array to isolate specific images
Extract Text From Document
Extracts text content from a PDF file and separates it into three categories:
Visual Descriptions — descriptions of scenes, objects, characters
Relevant Other Text — supporting information that may assist generation
Filtered / Irrelevant Text — removed noise, metadata, or non-useful content
This separation is useful when you want to feed only contextually relevant text into your generation nodes or LLM prompts.
Split Character Sheet
Detects and extracts individual views—such as front, side, and back—from a single character design sheet.
Automatically parses a design sheet to isolate different character poses into separate image outputs.
Works best with horizontal, vertical, or grid layouts where poses are clearly separated by empty space.
Outputs: Specific views for Front, Right, Left, and Back, plus an array of any "Other" detected angles (like 3/4 views).
Ideal for: Preparing character sheets for the Multi-View -> 3D node or organizing concept art.
Array Management Nodes
Break Images Array
Takes an image array and outputs up to four images separately.
Change Start Index to control which images are extracted
Useful when only specific images from a PDF or batch are needed
Ideal for directing selected images to multimodal or generation nodes
Create Images List
Collects up to 8 individual images and groups them into a single array output.
Use this when:
Preparing multi-image inputs for generation nodes
Organizing design references into one structured output
Combining image sets extracted from different parts of the workflow
Concatenate Image Arrays
Combines up to 4 image arrays into one unified array.
Ensures multiple sources of images appear in a single list
Output is a single array containing all items in order
Text & LLM Nodes
Combine Text
Merges multiple text inputs into one unified string.
Helps combine text blocks before sending them into generation nodes
Useful when you want to enforce a fixed prefix or order (e.g., system constraints + extracted text → final generation prompt)
Typical use case:
Merge a “generation instruction” with extracted document text
Feed the combined result into an image or 3D generation node
Text Generation (LLM)
LLM tool for controlled text generation.
Inputs:
System Prompt — defines the agent’s role and behavior
Text Input — any context or content to transform
Output: Single text result
Example use case:
You can input a concept art image and use Describe Image to extract the existing character list. A Text Generation (LLM) can then be instructed to return only the first character from that list. You may add an additional text input that defines how the multimodal node should isolate the selected character for 3D modeling. Using Combine Text, you merge the LLM-generated character description with your isolation instruction and feed the combined text directly into the multimodal node. This creates a controlled, image-conditioned prompt for generating a clean extraction of a single character ready for modeling.
Another example is providing an image of a full scene and using Describe Image to obtain a structured description of its spatial layout. You can then instruct the Text Generation (LLM) to act as a spatial-logic prompt generator that transforms these descriptions into a precise prompt for generating a 2D top-down plan of the same scene. By passing this prompt into a multimodal node, you obtain a clean plan abstraction. With this method, any input image produces a custom, image-specific plan prompt through the combined use of Describe Image, LLM processing, and Combine Text.
Simple Text Render
Generates a 2D image from a text string with automatic height adjustment to fit the content.
Allows for basic typography control, including font selection, size, line spacing, and color customization.
The output image width is fixed by the user, while the height scales dynamically based on how much text is provided.
Inputs: Input Text, Font Name (Arial/Nunito), Image Width, Font Size, and Background/Font Colors.
Ideal for: Creating labels, UI elements, or adding textual descriptions directly into your image processing pipeline.
Concatenate Images
Combines multiple images into a single image file by joining them side-by-side or top-to-bottom.
In Horizontal mode, all images are automatically scaled to match the height of the first image in the array.
In Vertical mode, all images are scaled to match the width of the first image in the array to maintain a uniform column.
Inputs: Input Images (Array), Direction (Horizontal/Vertical).
Ideal for: Creating comparison grids, before-and-after montages, or reassembling split character sheets
Utility Nodes allow these types of modular logic chains ; extracting structured information, refining or transforming it, and recombining it to produce workflows that adapt automatically to the input image while remaining deterministic and reusable across the Atlas platform.
Last updated