ComfyUI Launch Original API Node

ComfyUI has launched Native API Nodes which allows users to ** call multiple payment models API** directly in the workflow, e.g. Google Veo2, <a href=”/tags/openai.html’ | prepend: site.baseurl }}” class=”tag-link-inline”>OpenAI</a> GPT-4o image, Stability AI, Luma, Recraft, Pika 2.2, PixVerse, Ideogram etc. 11 model series, 65 nodes. This means that users can call different models, such as images, video generation, text-to-videos, in parallel to the same ComfyUI process** without exiting the interface and organizing the generation tasks.** In addition, ComfyUI supports:

All API node functions are ** optional** and the ComfyUI body is still ** completely free of charge**.

ComfyUI original API node?

ComfyUI used to be a graphical node-driven local image-generation tool, widely used in the Stable Diffusion community. However, this update introduced a completely new component: ** Original API Node: ** Allows users in ComfyUI** to call directly multiple mainstream APIs without leaving the interface or writing codes. This means: You can use a flexible mix of business models (image/video/multi-model)** in a workflow, unify scheduling and generation, and significantly improve creative efficiency.

The model currently supported is a look at API

ComfyUI Adds primary access support to the following 11 model series totals 65 nodes covering multi-model tasks such as image, video, text-to-video, image-to-video.

#

👉 用户通过拖放这些节点即可调用模型,无需单独登录每个平台,支持组合调用。

Details of how to use the workflow

# Use the process:

update ComfyUI or ComfyUI Desktop Login / Register account, purchase or bind API fractions of an existing platform Open: Workflow BrowseTemplates Image API / Video API Select template to run directly ✅ 你也可以使用“Bring Your Own API Key(自带 Key)”模式,绑定自己在其他平台已有的 API 凭证,例如你在 Pika、Luma、<a href=”/tags/openai.html’ | prepend: site.baseurl }}” class=”tag-link-inline”>OpenAI</a> 上已有订阅或额度。

#

When calling multiple external APIs (e.g. image+video+subtitles), ComfyUI automatically executes different nodes in parallel, significantly increasing overall generation speed. Examples of multi-model collaboration:

Luma Photon, plus real sense of light.

  • Scene + Prompt PixVerse Generate Video Snippets

  • Video + Description Pika Generate Animated White Camera

First introduction of VIDEO type native support

This is the first time that ComfyUI supports a “video generation” type of node (VIDEO type), meaning that:

A new upgrade of visual brands

In addition to technical upgrades, ComfyUI has officially launched a brand-new visual system:

  • Logo: consists of several modules that respond to the node graphical workflow;

  • ** fonts and colours: ** joined 90-year kinetic style + Y2K digital sense;

  • ** Expression of ideas: ** Retains the community spirit of “free, accessible, hacking”, emphasizing that instrumentality coexists with open creativity. Officially, “What we want to convey is that ComfyUI is still free and open, but it has also become a powerful tool that can really be used in the production process.”

What’s the point of summing up?

  1. ** Uniform multi-model access platform** Integration of multiple mainstream API calls in a workflow will no longer be required to switch platforms and manually upload download materials.
  2. ** Free body + optional commercial API** The ComfyUI main body is still open and free of charge, and the API node is used on demand. ** “Multi-modular generation tubes” for real creators** Images, videos, text, structure data, styles are integrated and it will be possible to add audio/3D nodes in the future.

Who’s fit for these functions?

Official presentation: https://blog.comfy.org/p/comfyui-native-api-nodes