Utility Nodes

Utility nodes often focus on models that can make generated outputs more stylistically consistent with both aesthetic and technical requirements of a production

Utility Nodes

Utility Nodes provide general-purpose functions that support all Atlas workflows. They are not tied to image or mesh generation; instead, they handle documents, text processing, arrays, and lightweight LLM operations.

Utility Nodes are essential for orchestrating complex workflows, preparing inputs, and structuring data for generation nodes.

Document Nodes

Extract Images From Document

This node processes an uploaded PDF document and extracts all images embedded within it.

  • Input: PDF file (e.g., a Game Design Document)

  • Output: Image Array

  • Each extracted image is returned as an element in the array.

You can use the resulting array with:

  • Find Images by Description node to pull specific images

  • Image Generation nodes that accept image arrays

  • Break Images Array to isolate specific images

Extract Text From Document

Extracts text content from a PDF file and separates it into three categories:

  • Visual Descriptions — descriptions of scenes, objects, characters

  • Relevant Other Text — supporting information that may assist generation

  • Filtered / Irrelevant Text — removed noise, metadata, or non-useful content

This separation is useful when you want to feed only contextually relevant text into your generation nodes or LLM prompts.

Split Character Sheet

Detects and extracts individual views—such as front, side, and back—from a single character design sheet.

  • Automatically parses a design sheet to isolate different character poses into separate image outputs.

  • Works best with horizontal, vertical, or grid layouts where poses are clearly separated by empty space.

  • Outputs: Specific views for Front, Right, Left, and Back, plus an array of any "Other" detected angles (like 3/4 views).

  • Ideal for: Preparing character sheets for the Multi-View -> 3D node or organizing concept art.

Array Management Nodes

Break Images Array

Takes an image array and outputs up to four images separately.

  • Change Start Index to control which images are extracted

  • Useful when only specific images from a PDF or batch are needed

  • Ideal for directing selected images to multimodal or generation nodes

Create Images List

Collects up to 8 individual images and groups them into a single array output.

Use this when:

  • Preparing multi-image inputs for generation nodes

  • Organizing design references into one structured output

  • Combining image sets extracted from different parts of the workflow

Concatenate Image Arrays

Combines up to 4 image arrays into one unified array.

  • Ensures multiple sources of images appear in a single list

  • Output is a single array containing all items in order

Text & LLM Nodes

Combine Text

Merges multiple text inputs into one unified string.

  • Helps combine text blocks before sending them into generation nodes

  • Useful when you want to enforce a fixed prefix or order (e.g., system constraints + extracted text → final generation prompt)

Typical use case:

  • Merge a “generation instruction” with extracted document text

  • Feed the combined result into an image or 3D generation node

Text Generation (LLM)

LLM tool for controlled text generation.

  • Inputs:

    • System Prompt — defines the agent’s role and behavior

    • Text Input — any context or content to transform

  • Output: Single text result

Example use case:

You can input a concept art image and use Describe Image to extract the existing character list. A Text Generation (LLM) can then be instructed to return only the first character from that list. You may add an additional text input that defines how the multimodal node should isolate the selected character for 3D modeling. Using Combine Text, you merge the LLM-generated character description with your isolation instruction and feed the combined text directly into the multimodal node. This creates a controlled, image-conditioned prompt for generating a clean extraction of a single character ready for modeling.

Another example is providing an image of a full scene and using Describe Image to obtain a structured description of its spatial layout. You can then instruct the Text Generation (LLM) to act as a spatial-logic prompt generator that transforms these descriptions into a precise prompt for generating a 2D top-down plan of the same scene. By passing this prompt into a multimodal node, you obtain a clean plan abstraction. With this method, any input image produces a custom, image-specific plan prompt through the combined use of Describe Image, LLM processing, and Combine Text.

Simple Text Render

Generates a 2D image from a text string with automatic height adjustment to fit the content.

  • Allows for basic typography control, including font selection, size, line spacing, and color customization.

  • The output image width is fixed by the user, while the height scales dynamically based on how much text is provided.

  • Inputs: Input Text, Font Name (Arial/Nunito), Image Width, Font Size, and Background/Font Colors.

  • Ideal for: Creating labels, UI elements, or adding textual descriptions directly into your image processing pipeline.

Concatenate Images

Combines multiple images into a single image file by joining them side-by-side or top-to-bottom.

  • In Horizontal mode, all images are automatically scaled to match the height of the first image in the array.

  • In Vertical mode, all images are scaled to match the width of the first image in the array to maintain a uniform column.

  • Inputs: Input Images (Array), Direction (Horizontal/Vertical).

  • Ideal for: Creating comparison grids, before-and-after montages, or reassembling split character sheets

Utility Nodes allow these types of modular logic chains ; extracting structured information, refining or transforming it, and recombining it to produce workflows that adapt automatically to the input image while remaining deterministic and reusable across the Atlas platform.

Last updated