Comfyui-zhenzhen
https://ai.t8star.cn/register?aff=dP7j Comfyui的各种平价API调用节点
Stars: 382
Comfyui-zhenzhen is a repository that provides various API nodes for generating images, videos, and text using different AI models. It includes nodes for models like Nano Banana, Veo3, Sora2, and more. The repository offers tools for image editing, video creation, and text generation, with options for customizing parameters and outputs. Users can access different AI models for specific tasks like image generation, video editing, and text creation.
README:
https://www.runninghub.ai/?inviteCode=rh-v1121
https://www.runninghub.cn/?inviteCode=rh-v1121
My favorite girl Go YounJung夸克网盘:https://pan.quark.cn/s/445806b40baa 谷歌盘:https://drive.google.com/file/d/1Ie384HYnA_nF5yOUqi57jm1WdeqpLiVW/view?usp=sharing
本站开设初衷是提供平价的API给粉丝朋友玩最新的海外模型,并非盈利目的,秉承这一理念,我们的价格毛利不到10%,去掉正常缴税和人工开发维护,服务器成本后几乎没有利润,所以并非盈利性质网站,没有任何议价空间,也不支持用于商业目的的二次开发,仅服务于粉丝朋友,望理解,每个月发票数量有限,需要自己承担所有税费5-25%
网站使用教程:https://ai.t8star.cn/about
API调用及开发教程:https://ai.t8star.cn/api-set
B站:https://www.bilibili.com/video/BV1TzspzpEFJ/
Youtube:https://www.youtube.com/watch?v=zwpjDCMCJOY
B站:https://www.bilibili.com/video/BV13MyPBzEjJ/
Youtube:https://www.youtube.com/watch?v=J6Zg7SrU_BE
Bilibili教程:https://www.bilibili.com/video/BV1g1ULBmEhA/
Youtube教程:https://www.youtube.com/watch?v=MASUQv_SKGs
B站:https://www.bilibili.com/video/BV1pRCABsE9y/
Youtube:https://www.youtube.com/watch?v=LvvD5HkKqHM
B站:https://www.bilibili.com/video/BV1pRCABsE9y/ Youtube:待更新
B站:https://www.bilibili.com/video/BV16MChB8ERT/
Youtube:https://www.youtube.com/watch?v=GsQnGBqRPnA
整合包V2:https://pan.quark.cn/s/86d82abd2f17
B站:https://www.bilibili.com/video/BV1nzUWB9EKx/
Youtube:https://www.youtube.com/watch?v=IZKLX1fyZIk
整合包V3:https://pan.quark.cn/s/635ee1b917b6
B站:https://www.bilibili.com/video/BV1zhUKBKEAp/
Youtube:https://www.youtube.com/watch?v=hRpdlEWCOws
B站:https://www.bilibili.com/video/BV1H3UuB2EJ3/
Youtube:https://www.youtube.com/watch?v=HLYNWk3B1ho
Bilibili教程:https://www.bilibili.com/video/BV1wUSmBfEhv/
Youtube教程:https://www.youtube.com/watch?v=_5CPM8vOi_s
Bilibili教程:https://www.bilibili.com/video/BV1zs2WBjEsh/
Youtube教程:https://www.youtube.com/watch?v=KB5NDfQ65Ug
Bilibili教程:https://www.bilibili.com/video/BV1SmmnBoEDz/
Youtube教程:https://www.youtube.com/watch?v=P3zMzPLSLT4
如果报错了,请查看上面的代码,不一定都是你的问题,或者服务器问题,尤其SORA2,大部分情况是OPENAI的问题,当你发生500报错了,请在跑一次就可以了!
由于NanoBanana Pro 4K的图像文件过大,在跨境下载时候容易导致失败,新增OSS设置(对象存储服务)
登录网站-令牌-你使用的APIKEY-编辑-拉到底部-选择CN或者US的OSS即可解决问题,如果报错也可以在comfyui终端检查最后报错代码,其中有下载地址,可以下载图片!
这几天洛杉矶办事处网络出现一些问题,由于亚太地区海底光缆被切断导致我们骨干网络服务供应商NTT网络性能下降,目前该服务商已暂时停止服务。所波及影响是无论打开我们贞贞的AI工坊或者连通,请求API都需要开启魔法,望周知(修复时间未知) 部分人完全是因为用的Vpn软件问题,请用常规软件!开Tun模式 开源免费版仓库:https://github.com/2dust/v2rayN/releases 下载Windows版本! 另外360防火墙这些垃圾软件也可能导致这个问题!请关闭!包括有些人的路由器设置!
更新Grok3VideoApi节点,目前支持15S视频!漫剧,真人短剧必备!
支持gemini-3.1-flash-image-preview模型,新增节点Zhenzhen Gemini 3.1 Flash Image Edit S2A 更新工作流zhenzhen-gemini-3.1-flash-image-preview_edit_s2a.json 更新工作流zhenzhen-gemini-3.1-flash-image-preview_t2i_s2a.json
已支持 glm-5模型,目前世界级sota存在,支持缓存计费,sora-2模型default 分组暂时恢复可用,sora-2 价格上升至 0.5 一次/15秒
新增sora-2-vip模型,单词1.25 只支持10S,分组选择default,生成成功率90%以上,生成时间150秒左右 新增工作流:zhenzhen-sora2-2-vip-ti2v-recommend.json
支持grok video新模型的10秒功能,增加分辨率选项,支持文生视频和图生视频,支持中文语言,删除老的工作流,新增2个新的工作流 zhenzhen-grok3-video-t2v-v2.json zhenzhen-grok3-video-i2v-v2.json
更新sora2_new节点 进一步加强稳定性
目前sora2进展, 1、目前sora2-vip分组基本可用,但是成本上涨即将涨价 2、sora2的default分组已经开始降低失败率,即将稳定,成本上涨即将涨价 3、sora2_new模型,独立API仅提供漫剧即将违约的朋友,成本价1000条200元,网站公告有联系方式,需要预约,数量有限,纯帮忙
sora2平替: gork视频模型已经升级,支持多镜头,效果比Sora2还要好,支持直接真人模式,无限并发,NSFW我们已经过滤,支持生成10s,有声音。是sora目前比较合适的平替。完美匹配中文语音,效果炸裂!!如果不会调用的可以看我的comfyui节点,扔给AI!模型名字grok-video-3 视频效果可以看贞贞工坊官网-左侧教程-最上方
Sora2基本搞定,测了一天了,很稳定,模型请用:sora-2, 分组切换成 sora-vip,消费积分0.33334/次,请自行尝试,,15S,请自行尝试
目前可以用的是Default分组的sora-2-pro模型,支持HD和25S,以及sora-vip分组的sora-2模型
sora2-new模型稳定版已支持自动去水印(非并发comfyui版,支持RH使用) 这几天洛杉矶办事处网络出现一些问题,由于亚太地区海底光缆别切断导致我们骨干网络服务供应商NTT网络性能下降,目前该服务商已暂时停止服务。所波及影响是无论打开我们贞贞的AI工坊或者联通API都需要开启魔法,望周知!
新增了sora2_new节点,需要单独的APIKEY,因为共同的最近风控太厉害了,如果有需要的B站联系我。。稳定 新增了veo3.1-fast模型的支持 新增sora2 new工作流,和原本的APIKEY不通用
暂时取消veo3.1-fast模型,有点问题,等处理后添加,重写veo3.1节点,务必更新!
新增veo3.1-4k工作流
sora2-character角色获取节点新增from_task参数,创建角色可以通过taskid创建
更新支持veo3.1-fast,4k,pro 4k等模型及节点支持,fast 基准0.2 4K 基准0.5 Pro 4K 基准2,工作流不变,只需要选择新模型
更新vidu q2多参工作流:zhenzhen-vidu-q2-multi-ref-i2v.json
Api Setting节点废弃2个不用的接口地址(US,HK)
更新支持minimax hailuo 2.3 I2V图生视频及工作流
更新支持Flux2 max图像编辑及工作流
更新支持Mj niji 7及工作流,注意这个目前不支持用apisetting节点,我下个版本更新,直接输入KEY即可
网站教程顶部增加基础报错教程以及企鹅无限并发整合包
Claude 4.5 opus Default渠道恢复 Veo3.1 恢复 Grok vieo 3 恢复 Sora2正在继续扩容,目前出现超时请重试,我们正在不断扩容!
目前由于openai官方问题,sora2的API大批出现问题,等待官方修复,并非我们问题
升级gpt-image-1-edit节点,支持最多10图输入和10 mask(Mask我没测试) 新增2个gpt-image-1-edit工作流
新增gemini-3-flash-preview模型 昨天谷歌大香蕉2 Default分组基本一点问题没有,大家可以继续用便宜的分组,优质分组做紧急备用
更新gpt-image-1和gpt-image-1-edit节点,支持gpt-image-1.5模型 推荐优质分组基准0.12,Default分组超级慢,不推荐,非异步,目前测试优质分组大概200秒左右出图
香蕉2模型的优质分组价格阶梯:1k,2k都是0.42算力额度, 4k单独做了价格:0.555算力额度。默认分组价格保持不变。
重要通知, 谷歌Nano Banana 2(pro) 大香蕉官方低价渠道并发削减导致卡顿,后台排队阻塞严重,现提升 Gemini 优质分组价格,使用高等级 Vetrex 账号,以维持稳定可用;default组得等谷歌调整恢复并发,可用性较差,急的建议尝试调整令牌分组,网站左侧令牌,找到对应的APIKEY编辑,添加gemini优质分组
新增节点wan2_6_API,支持官方百炼的APIKEY,暂时贞贞网站还没上,我测试用的,支持文生和图生视频,工作流已更新到workflow目录下 llm_api节点重写视频识别部分,目前已正常支持视频反推,工作流已更新到workflow目录下
修复代码冗余导致的节点加载异常
sora2_character新增from_task参数输入框:可以直接把sora2生成的真人视频的task id直接输入进去创建角色。
今天openai服务器问题导致sora2一直失败(应该是GPT5.2太火爆),导致的负载不够,并非我们问题,SORA2只能多抽卡或者等官方恢复后使用,并非我们问题
新增节点zhenzhen_llm_api,支持图像反推,文本扩写,支持Openai兼容格式,可用网站的模型,如需反推的话需要标签带识图的模型 模型选择位置-登录网站-右上角模型价格-搜索或者查找你要的模型,注意标签哦!
Sora2 Pro 15S HD版和25S已经恢复使用,节点不变,基准2.5元
Nano-banana-2(Pro) 由于官方并发大幅缩减,目前正在大量增加并发,成本上涨,故NanoBanana Pro价格从0.15上调至0.2基准,gemini 3 pro image preview暂时不变,价格已经生效,望周知
新增Grok3-video节点,支持最新GORK文生视频和图生视频,支持音画同步,支持中文配音,已过滤nsfw内容,合规使用!新增对应的2个工作流 今天下午谷歌砍了并发,导致目前一直分组上限,目前在逐步处理,推荐先用优质分组
新增工作流zhenzhen-Nano-Banana-2-Edit-S2A.json,同步转异步,支持异步任务,可在网站后台异步任务查询(nano banana pro)
新增nano_banana2_edit_S2A节点,支持同步转异步模式,任务可在异步任务中查询了!同时可以通过任务ID回溯下载失败的图片URL地址!另外做了一些异常处理
修复veo3.1 component模型问题 新增节点:Zhenzhen_Z_Image_Turbo 工作流晚上回去发
新增Zhenzhen Doubao Seedream4.5节点,可使用豆包4.5,出图速度比大香蕉2快多了,人物一致性好过大香蕉,基准0.15,支持2K和4K 新增zhenzhen-doubao-seedream4_5_image_edit.json工作流
nano_banana2-edit节点新增模型nano-banana-2-2k,nano-banana-2-4k,如果用API调用时候,image_size参数不生效,主要用于由于谷歌问题导致的2K,4K出图变1K问题的备用方案,作为备用模型
上架可灵视频O1模型,Z-image-tubro以及阿里的各种视频模型
以下2个新模型扣费方式不太一样,使用前请注意下,因为都是官方渠道
更新vidu以及Flux2 pro和Flux2 Flex新工作流
Vidu的扣费方式是按照0.04*官方的Cost
flux-2-pro模型,此模型按量付费,不同分辨率不同价格
官方文档:
https://docs.bfl.ai/api-reference/tasks/generate-or-edit-an-image-with-flux2-[pro]
价格计算器:
上线 Claude 最新模型 claude-opus-4-5-20251101,目前最强的推理模型之一,上架MinMax系列音频模型,上架Flux2 Pro和Flux2 Flex模型,上架Vidu多参,文生,图生,首尾帧模型
SORA2 15S和10S调用变慢的问题,目前OPENAI又改了,所以速度变慢了,在优化,但是速度应该回不到之前速度,每天晚上NANO BANANA PRO速度变慢时候,建议把BASEURL改成US,白天正常的情况下用主站
新增Flux2相关节点,Flux_2_Flex,Flux_2_Pro 新增Vidu相关节点,vidu_img2video,vidu_text2video,vidu_ref2video,vidu_start-end2video
工作流晚上更新
Nano Banana 2模型后端进行了调整,基本修复了出白图的问题,调整了格式,可以尝试使用Default或者和优质分组一起用
Nano Banana 2 Edit节点新增AUTO模式,自动跟随图片尺寸比例,已默认
注意,目前令牌分组default负载过大,4K容易出白图,推荐2K 如需4K,在贞贞AI工坊网站左侧令牌-找到自己的令牌编辑,删除Default分组,添加优质分组,成功率大大提高 报错429和报错500都是谷歌负载上限,并不扣费,如果遇到的是提示无法处理图片并扣费,这时候请去网站的日志里,找到对应任务点进去,实际已经生成完毕,目前在排查这个问题
Nano Banana 2模型4K价格下调到基准0.15元(rmb) Veo3.1及component模型价格从2.025下调到0.3元(rmb) Veo3.1 Pro模型价格从3元下调到1元(rmb)
新增节点:nano_banana2_edit
支持Nano Banana 2多图编辑以及2K,4K高清版
增加工作流:Nano-Banana-2-Edit-T2I-4K.json
增加工作流:Nano-Banana-2-Edit-Multi-I2I-4K.json
可选分辨率以及1K,2K,4K,暂时都是0.2积分
新增Nano-Banana-2模型
目前支持2个模型Gemini-3-Pro-Image-Preview 和 Nano-Banana-2
区别:
Gemini-3-Pro-Image-Preview,适配官方格式级openai chat格式 nano-banan-2:支持chat,支持Openai兼容格式,传送支持2K,4K,API调用参数:mage_size="4K",支持绘图接口同步模式以及异步模式
节点也更新了,但是似乎4K没生效,白天看下在修复
支持nanobanana 2,直接用原本的nanobnana工作流即可,更换下模型! 模型名字gemini-3-pro-image-preview 增加工作流:zhenzhen-nano-banana-V2(Gemini-3-pro-image-preview).json 目前是0.2积分/图
增加2个新工作流,分别是Sora2角色创建节点,以及多角色@调用加图片参考混合工作流
zhenzhen-sora2-character-create.json --角色创建,创建将花费0.01积分,只需创建一次,可以无线调用
zhenzhen-sora2-multi-character.json --多角色@调用加图片参考混合
创建我们可以通过图床的URL获取,但是很多图床不支持VIDEO,另外有些哪怕可以上传SORA2也不支持,所以推荐用下面方式:
首先先用Sora2跑一次视频,跑出来的角色的response的MP4地址复制到新节点的URL
常规操作:
URL部分需要用图床,但是不是每个图床都支持,最好的办法, 先用我们的API生成一个SORA2视频,他会有个MP4地址
生成的usename保存下来,后续就可以直接@使用了,支持多角色
比如:https://midjourney-plus.oss-us-west-1.aliyuncs.com/sora/39cf59c1-354f-4ba3-88d6-5abebc14c0b8.mp4
让后复制下即可,就能获得新的USERNAME
timestamps代表获取URL视频的其中几秒中的角色,1,3意思就是1-3秒,以此类推
参考工作流是2角色@加一个参考背景,可支持多角色客串艾特以及多图参考,当然太多肯定也会影响一致性,请适量!
增加3个新工作流,分别是Suno V5歌曲翻唱,Suno V5歌曲续写,Suno V5歌词读取,3个文件已更新到workflow文件夹下,一次出两首歌,如果发现两首歌一样的话,只需要去贞贞AI工坊网页版-左侧的异步任务,点开任务ID,里面下载MP3即可
新增Comfly Sora2 Character节点,使用方式是把生成后的username用于放在提示词中 @username 直接调用。可同时使用多个角色客串调用。需要先在sora2中认证创建角色,或者用别人已认证创建过的角色,支持真人!
统一所有的节点分组
sora2新增创建角色 API,创建角色,后续可在prompt 里 @ 调用,可以同时调用多个角色,调用方式:https://gpt-best.apifox.cn/api-374618722
新增VIDU接口
节点后续更新
新增sora2-pro优质分组,网站显示有问题,不是次数,是按秒收费,务必注意!还是比较贵的,只有4秒8秒12秒,原本的也能用,但是奥特曼改了每个账号的次数,所以每天次数有限
插件更新suno_cover,suno_upload_extend,suno_upload:节点`:
suno_upload:上传自己的音频文件,获取clip_ip,用于suno_cover和suno_upload_extend节点。上传音频时长必须在6s-60s内。
suno_cover:音乐翻版\修改风格,输入clip_ip(可以是自己上传的,或者在平台生成的音乐的。PS:自己上传的有可能因为跨账号问题不生效。)
suno_upload_extend:音频续写,输入clip_ip(可以是自己上传的,或者在平台生成的音乐的)
示例工作流后续更新
增加新节点sora2_openai 是官方格式的节点,暂时只支持图生视频,不支持文生视频
所有Sora2节点新增参数,防止生成视频泄露,已经内置,不需要手动设置
修正最新comfyui内核节点加载报错的问题
RunningHub上所有API相关工作流均已更新,如果发现有漏网之鱼请B站或者youtube留言发链接,我会及时处理
更新sora2-chat节点,速度更快,目前只支持横版,竖版15S非HD视频!不支持HD和25S,更稳定
所有工作流全部做了更新,请下载新工作流,搭建方式有所区别
新增sora2-chat对应工作流
另外Sora2-chat节点,生成后可以下载GIF以及视频,在Comfyui节点会生成如下链接,请从控制台中查找
然后链接复制到浏览器打开,如下也可以下载和分享
新增节点api setting 由于昨天网站被攻击,昨天半夜已经修复,防止再出现类似问题,增加多个可选接口,目前有1个主站,一个美国站,一个香港站,还有个自定义IP接口,RH的工作流需要晚上更新,本地直接改下节点即可
修复上个版本NananaBanana图像编辑比例不生效问题 新增12个包括nanobanana以及veo3,veo3.1的工作流
veo3特价模型基准1.5RMB/8S,模型名字: veo3 文生视频 veo3-fast 文生视频 veo3-fast-frames 图生视频
Veo3 Pro模型3RMB/8S,模型名字: veo3-pro 文生视频 veo3-pro-frames 图生视频 高质量模式,价格也高一点
Veo3.1 特价模型基准1.5RMB/8S,模型名字: veo3.1 文生视频,图生视频 veo3.1-components 文生视频,图生视频,多参1-2张 注意veo3.1-components,支持多图参考,Veo3.1支持首尾帧
Veo3.1 Pro模型基准3RMB/8S,模型名字: veo3.1-pro 文生视频,图生视频,首尾帧 高质量模式,价格也高一点
注意模型价格可能根据情况调整,实际价格以https://ai.t8star.cn/register?aff=dP7j 为准
Nano-banana 由于长期风控原因,将于 10.20 日 中午12点涨价至 0.08/次; 官方模型(gemini-2.5-flash-image)价格不变 gemini-2.5-flash-image和gemini-2.5-flash-image-preview模型有概率不出图,只支持chat格式
nano-banana节点已更新,只有edit节点才支持尺寸选择。
Veo3.1节点更新,支持新模型veo3.1,价格不变
工作流一会更新
目前模型分为sora-2和sora-2-pro版
sora-2 偶发性出水印,价格基准0.1RMB,不支持HD,可选10S和15S,支持横版竖版,画面会有闪烁 sora-2 Pro,无水印,基准价格1.7RMB,10S和15S支持HD,25S不支持HD,支持横版竖版,无闪烁画面,HD非常清晰,非HD一般清晰
务必注意25S不要和HD同时开!
25S 大概需要30分钟 15S HD大概需要20分钟 10S 非HD大概5分钟左右 15S 非HD大概10分钟左右
由于并发压力不同,以上时间仅供参考
更新了6个最终版工作流,可以直接使用
B站教程:https://www.bilibili.com/video/BV1m3WszsEn7/ Youtube教程:https://www.youtube.com/watch?v=bc8b5ZxPvJE
修复了sora-2-pro模型,其中10S和15S支持HD同时开启,25S不支持HD,务必注意!基准都是1.7RMB,无论是不是HD或者是不是25S,25S目前测试下来需要1400秒左右,非常久,务必注意 sora-2模型,是普通版模型,可以输出15S,支持横版竖版,时间预计5-10分钟,基准价格0.1RMB 修改了后台的超时等待时间,从1200秒提高到了3000秒,节点也对应修复
再说一遍,务必注意,HD不支持25S!!!
更新了工作流,删除老的工作流 目前15S没问题,时间有所延长 25S和HD有点问题,暂时先不要用,白天排查
注意,Sora2更新节点后15S不要开HD,否则会报错!25S才能开HD!
修改Sora2的最大重试次数
支持Sora2 HD 25S 以及 普通版 15S,无水印,节点已更新,工作流重新拉一下节点即可 Veo3.1已支持,节点明天更新 NanoBanan自定义分辨率已更新,节点明天更新
由于sora2 HD 15S非常需要时间(官方也很慢),将重试时间从120S-调整到180S,目前测试下来大部分在150S-160S时候成功,偶尔也有180S没成功情况 节点超时不代表一定失败,请从贞贞AI工坊-左侧异步任务,点进去可以看看到,点对应任务的蓝色数字链接,从里面下载MP4即可,如果超时会提示超时,超时失败会退费 Sora普通版0.1积分,SORA HD 15S 1.7积分
删除2个老的SORA2工作流,老版本HD和15S模型下架 sora_video2-portrait-15s 模型下架 sora_video2-landscape-15s 模型下架 sora_video2-portrait-hd-15s 模型下架 sora_video2-landscape-hd-15s 模型下架
新的模型请用:sora-2-pro 工作流:sora2-plus-new-15shd.json
这个版本质量非常高,但是跑起来非常慢,15S差不多10分钟,HD还要单独8分钟,也就是跑一次至少20分钟左右,且有失败概率,如果Comfyui超时了,请去网站查看下异步任务,看看原因,正常失败会退积分。
Comfly_sora2:节点,新增sora-2-pro模型,目前基准1.7,目前出视频速度较慢,无水印,支持15秒以及HD,工作流同步更新,workflow目录下sora2-plus-new-15shd.json
Comfly_sora2:节点: 新增sora2视频模型节点,新节点支持直接video combine了,不用再复制链接了!目前无水印,生成最多10s普通画质视频,hd和15s暂时无法使用请知晓。
更新新节点:OpenAi sora api plus:支持4个新模型,支持15秒时长
上面2个1080P,下面2个是HD,国庆期间依然粉丝价格0.01人民币一个视频
sora_video2-portrait-15s
sora_video2-landscape-15s
sora_video2-portrait-hd-15s
sora_video2-landscape-hd-15s
目前国庆期间依然15S依然是0.01元RMB,给大家玩,祝大家中秋快乐!
更新新工作流:Sora2-Plus-15s-HD.json,15秒专属工作流,支持HD,支持横版及竖版
新版节点可以用以下5个模型,支持横版,复制到节点的model上即可,国庆期间PRO模型也是0.01元RMB,国庆礼物,祝大家国庆玩得开心
竖版高清模型:sora_video2-portrait-hd 横版高清模型:sora_video2-landscape-hd 新版竖版模型:sora_video2-portrait 新版横版模型:sora_video2-landscape 老板竖版模型:sora_video2 更新节点以及示例工作流--Sora2-hd.json
Sora2 Api目前并发较高,如反复出现500代码且报错代码为401,通常是这个APIKEY首先,重新新建一个令牌一般即可解决
20251002:
新增Sora2节点,国庆期间每个10秒视频仅0.01元,目前保存视频节点有点问题,保存video_url,然后复制到浏览器保存视频即可
20250924:
Comfly_suno:节点: 新增v5模型
Comfly_suno:节点: 新增Comfly_suno_description,Comfly_suno_lyrics,Comfly_suno_custom三个节点
简单描述生成歌曲,生成歌词,自定义生成歌曲三个节点。
Comfly_Doubao_Seedream_4节点: 节点新增自定义尺寸。在aspect_ratio选择Custom,然后可以在width和height自定义。
Comfly_Googel_Veo3:节点: Veo 模型大幅降价,文生视频支持设置横、竖屏
Comfly_Doubao_Seedream_4节点: 新增节点:"Comfly Doubao Seedream4.0
Comfly_gpt_image_1_edit节点: 参数新增input fidelity,partial_images参数
Comfly_nano_banana_edit节点: 新增节点Comfly_nano_banana_edit,这个可以选择生成图片的尺寸,模型只能是:nano-banana
文生图下尺寸才能生效,图生图不生效。
Comfly_MiniMax_video节点: 新增节点Comfly_MiniMax_video,支持海螺ai全部视频模型,支持最新首尾帧。
具体模型能力和参数选择请查看官方文档,避免使用错误:
https://platform.minimaxi.com/document/video_generation?key=66d1439376e52fcee2853049
目前官方返无图的可能性比较高,所以需要你开魔法,并且节点在美国(我测试这样的情况基本没有问题,有问题加群)
Comfly_nano_banana_fal节点: 新增节点Comfly_nano_banana_fal,这个可以生成1到4张图片,nano-banana为文生图模型。
nano-banana/edit为图生图模型(图生图模型会产生额外的图片上传费用,具体可以看网站日志,在网站异步任务也可查看任务信息)
Comfly_nano_banana节点: 新增模型nano-banana选项,这个模型不容易被识别成对话模型,
Comfly_nano_banana节点: 新增节点:Comfly_nano_banana(文生图,图生图,支持多图参考编辑),
谷歌最强编辑模型:gemini-2.5-flash-image-preview,
有默认和gemini优质两个分组。价格比官方便宜很多。可以在cherrystudio里面的newapi供应商填写我的api中转站调用模型使用。
qwen image_edit节点: 新增千问图片编辑节点:Comfly_qwen_image_edit,价格0.1.
可以自定义尺寸(size选择Custom后,在Custom_size输入分辨率即可,例如1280x720)。
num_images生成图片数量是1到4张,注意api计算是按照图片张数来的,生成越多,api消费就多。
doubao节点: 新增节点:Comfly_Doubao_Seedream和Comfly_Doubao_Seededit都是3.0模型
qwen image节点: 新增千问绘图节点:Comfly_qwen_image,价格全网最低~
可以自定义尺寸(size选择Custom后,在Custom_size输入分辨率即可,例如1280x720)。
num_images生成图片数量是1到4张,注意api计算是按照图片张数来的,生成越多,api消费就多。
mj 换脸节点: 新增mj换脸节点:Comfly_Mj_swap_face,修复mju,mjv节点bug。
kling 可灵节点: 新增可灵多图参考视频节点:Comfly_kling_multi_image2video,最多支持4个参考图,只支持1.6模型。
新增2.1模型选择。
mj video延长节点: 新增mj视频延长节点:Comfly_mj_video_extend,一次生成4个视频,按次收费。
task id是接入上一次生成视频的task id 输出内容。 index 是选择延长上一次生成的4个视频里面的哪一个做为延迟,范围是0,1,2,3,对应的是第一,二,三,四视频 视频最多延长4次,一次延长4s。
mj video节点: 新增mj视频节点:Comfly_mj_video,一次生成4个视频,按次收费。
20250716:删除了Comfly_kling_videoPreview节点,视频节点的video输出接口可以直接连接comfyui本体的save video节点。
20250714:
Googel veo3节点: veo3谷歌视频,新增veo3-fast-frames模型,图生视频
Googel veo3节点:
新增Comfly_Googel_Veo3节点,文生视频模型:veo3,veo3-fast,veo3-pro。图生视频模型:veo3-pro-frames。 enhance_prompt开关: 是否优化提示词,一般是false;由于 veo 只支持英文提示词,所以如果需要中文自动转成英文提示词,可以开启此开关。 目前4个模型都是自动生成带音效的。无法手动关闭,并且不支持选择生成视频尺寸,默认都是生成横幅视频。
Flux节点: Comfly_Flux_Kontext,Comfly_Flux_Kontext_Edit两个节点新增flux-kontext-dev模型
Flux节点: 新增bfl官方节点:Comfly_Flux_Kontext_bfl节点,价格不变
Flux节点: Comfly_Flux_Kontext_Edit节点支持设置出图数量(1-4张范围),这个节点不会消耗上传图片费用,直接传入图片即可,
跟Comfly_Flux_Kontext一样,就是上传图片不会扣费,图片输入支持base64图片编码格式,可以做为稳定性的备用节点。
Flux节点: Comfly_Flux_Kontext节点支持设置出图数量(1-4张范围),支持多图输入。
已经支持对上一次生成的图片再次提示词编辑(但只有当出土数量选择1时才可以使用这个。
Jimeng即梦视频节点: 新增ComflyJimengVideoApi节点。即梦视频,按次收费,5s是0.6元,10s是1.2元。
Flux节点: 新增Comfly_Flux_Kontext节点,支持:flux-kontext-pro和flux-kontext-max模型,按次收费:pro模型大约0.096元,max大约0.192元,比官方便宜很多。
20250518:
Kling节点: 可灵节点新增kling-v2-master的可灵2.0模型。价格很贵,按需使用。
20250429:
Chatgpt节点: Comfly_gpt_image_1_edit新增chats输出口,输出多轮对话。
新增clear_chats,当为Ture的时候,只能image输入什么图片修改什么图片,不支持显示上下文对话。
当为Flase的时候,支持对上一次生成的图片进行二次修改。支持显示上下文对话。并且支持多图模式下新增图片参考。
20250425:
Chatgpt节点:
新增Comfly_gpt_image_1和Comfly_gpt_image_1_edit官方gpt_image_1模型api接口节点。
模型名都是gpt_image_1,区别只是分组不同:
一共四个分组:default默认分组为官方逆向,价格便宜,缺点就是不稳定,速度慢。按次收费。不支持额外参数选择。这个分组的apikey只能用于ComflyChatGPTApi节点。
其他三个组都是官方api组,最优惠的目前是ssvip组。分组需要再令牌里面去修改选择。这3个官方分组优点就是速度快,稳定性高。支持官方参数调整。 缺点就是贵,但是也比官方便宜。大家可以按照自己的情况选择。这3个分组的令牌的apikey只能用在下面2个新节点上面!!!
-
Comfly_gpt_image_1 节点:文生图,有耕读参数调整,支持调整生图限制为low。
-
Comfly_gpt_image_1_edit 节点:图生图,支持mask遮罩,支持多图参考。
20250424:
Chatgpt节点: ComflyChatGPTApi节点新增官方gpt-image-1,按次计费 0.06,
旧版的gpt4o-image,gpt4o-image-vip,sora_image, sora_image-vip可以做为备选。首选gpt-image-1。
jimeng即梦节点: 即梦的ComflyJimengApi节点新增参考图生成图片,image url图片链接参考生成图片。
注意:参考图生成图片会额外消耗上传图片的token费用(具体根据你图片大小来,大部分都是0.000几到0.00几元不等。图片链接有时效性,不做长期储存),
这个只适用于你没有image url图片链接的前提下使用。
如果你有image url图片链接,就直接填写在image url里面既可以。
Chatgpt节点: ComflyChatGPTApi节点新增chats输出口,输出多轮对话。
新增clear_chats,当为Ture的时候,只能image输入什么图片修改什么图片,不支持显示上下文对话。
当为Flase的时候,支持对上一次生成的图片进行二次修改。支持显示上下文对话。
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for Comfyui-zhenzhen
Similar Open Source Tools
Comfyui-zhenzhen
Comfyui-zhenzhen is a repository that provides various API nodes for generating images, videos, and text using different AI models. It includes nodes for models like Nano Banana, Veo3, Sora2, and more. The repository offers tools for image editing, video creation, and text generation, with options for customizing parameters and outputs. Users can access different AI models for specific tasks like image generation, video editing, and text creation.
AI-YinMei
AI-YinMei is an AI virtual anchor Vtuber development tool (N card version). It supports fastgpt knowledge base chat dialogue, a complete set of solutions for LLM large language models: [fastgpt] + [one-api] + [Xinference], supports docking bilibili live broadcast barrage reply and entering live broadcast welcome speech, supports Microsoft edge-tts speech synthesis, supports Bert-VITS2 speech synthesis, supports GPT-SoVITS speech synthesis, supports expression control Vtuber Studio, supports painting stable-diffusion-webui output OBS live broadcast room, supports painting picture pornography public-NSFW-y-distinguish, supports search and image search service duckduckgo (requires magic Internet access), supports image search service Baidu image search (no magic Internet access), supports AI reply chat box [html plug-in], supports AI singing Auto-Convert-Music, supports playlist [html plug-in], supports dancing function, supports expression video playback, supports head touching action, supports gift smashing action, supports singing automatic start dancing function, chat and singing automatic cycle swing action, supports multi scene switching, background music switching, day and night automatic switching scene, supports open singing and painting, let AI automatically judge the content.
rime_wanxiang_pro
Rime Wanxiang Pro is an enhanced version of Wanxiang, supporting the 9, 14, and 18-key layouts. It features a pinyin library with optimized word and language models, supporting accurate sentence output with tones. The tool also allows for mixed Chinese and English input, offering various usage scenarios. Users can customize their input method by selecting different decoding and auxiliary code rules, enabling flexible combinations of pinyin and auxiliary codes. The tool simplifies the complex configuration of Rime and provides a unified word library for multiple input methods, enhancing input efficiency and user experience.
ESP32_AI_LLM
ESP32_AI_LLM is a project that uses ESP32 to connect to Xunfei Xinghuo, Dou Bao, and Tongyi Qianwen large models to achieve voice chat functions, supporting online voice wake-up, continuous conversation, music playback, and real-time display of conversation content on an external screen. The project requires specific hardware components and provides functionalities such as voice wake-up, voice conversation, convenient network configuration, music playback, volume adjustment, LED control, model switching, and screen display. Users can deploy the project by setting up Xunfei services, cloning the repository, configuring necessary parameters, installing drivers, compiling, and burning the code.
ZcChat
ZcChat is an AI desktop pet suitable for Galgame characters, featuring long-term memory, expressive actions, control over the computer, and voice functions. It utilizes Letta for AI long-term memory, Galgame-style character illustrations for more actions and expressions, and voice interaction with support for various voice synthesis tools like Vits. Users can configure characters, install Letta, set up voice synthesis and input, and control the pet to interact with the computer. The tool enhances visual and auditory experiences for users interested in AI desktop pets.
go-stock
Go-stock is a tool for analyzing stock market data using the Go programming language. It provides functionalities for fetching stock data, performing technical analysis, and visualizing trends. With Go-stock, users can easily retrieve historical stock prices, calculate moving averages, and plot candlestick charts. This tool is designed to help investors and traders make informed decisions based on data-driven insights.
FisherAI
FisherAI is a Chrome extension designed to improve learning efficiency. It supports automatic summarization, web and video translation, multi-turn dialogue, and various large language models such as gpt/azure/gemini/deepseek/mistral/groq/yi/moonshot. Users can enjoy flexible and powerful AI tools with FisherAI.
unity-AI-Chat-Toolkit
The Unity-AI-Chat-Toolkit is a toolset for Unity developers to quickly implement AI chat-related functions. Currently, this library includes code implementations for API calls to large language models such as ChatGPT, RKV, and ChatGLM, as well as web API access to Microsoft Azure and Baidu AI for speech synthesis and speech recognition. With this library, we can quickly implement cross-platform applications on Unity.
claude-pro
Claude Pro is a powerful AI conversational model that excels in handling complex instructions, understanding context, and generating natural text. It is considered a top alternative to ChatGPT Plus, offering high-quality content with almost no AI traces. The article provides detailed information on what Claude is, how to access it in China, how to register, and how to subscribe using a foreign credit card. It also covers topics like using a stable VPN, obtaining a foreign virtual credit card, and a foreign phone number for registration. The process of purchasing a Claude Pro account in China is explained step by step, emphasizing the importance of following the platform's policies to avoid account suspension.
nonebot_plugin_naturel_gpt
NoneBotPluginNaturelGPT is a plugin for NoneBot that enhances the GPT chat AI with more human-like characteristics. It supports multiple customizable personalities, preset sharing, and various features to improve chat interactions. Users can create personalized chat experiences, enable context-aware conversations, and benefit from features like long-term memory, user-specific impressions, and data persistence. The plugin also allows for personality switching, custom trigger words, content blocking, and more. It offers extensive capabilities for enhancing chat interactions and enabling AI to actively participate in conversations.
LLMForEverybody
LLMForEverybody is a comprehensive repository covering various aspects of large language models (LLMs) including pre-training, architecture, optimizers, activation functions, attention mechanisms, tokenization, parallel strategies, training frameworks, deployment, fine-tuning, quantization, GPU parallelism, prompt engineering, agent design, RAG architecture, enterprise deployment challenges, evaluation metrics, and current hot topics in the field. It provides detailed explanations, tutorials, and insights into the workings and applications of LLMs, making it a valuable resource for researchers, developers, and enthusiasts interested in understanding and working with large language models.
VideoCaptioner
VideoCaptioner is a video subtitle processing assistant based on a large language model (LLM), supporting speech recognition, subtitle segmentation, optimization, translation, and full-process handling. It is user-friendly and does not require high configuration, supporting both network calls and local offline (GPU-enabled) speech recognition. It utilizes a large language model for intelligent subtitle segmentation, correction, and translation, providing stunning subtitles for videos. The tool offers features such as accurate subtitle generation without GPU, intelligent segmentation and sentence splitting based on LLM, AI subtitle optimization and translation, batch video subtitle synthesis, intuitive subtitle editing interface with real-time preview and quick editing, and low model token consumption with built-in basic LLM model for easy use.
LLMLanding
LLMLanding is a repository focused on practical implementation of large models, covering topics from theory to practice. It provides a structured learning path for training large models, including specific tasks like training 1B-scale models, exploring SFT, and working on specialized tasks such as code generation, NLP tasks, and domain-specific fine-tuning. The repository emphasizes a dual learning approach: quickly applying existing tools for immediate output benefits and delving into foundational concepts for long-term understanding. It offers detailed resources and pathways for in-depth learning based on individual preferences and goals, combining theory with practical application to avoid overwhelm and ensure sustained learning progress.
For similar tasks
Comfyui-zhenzhen
Comfyui-zhenzhen is a repository that provides various API nodes for generating images, videos, and text using different AI models. It includes nodes for models like Nano Banana, Veo3, Sora2, and more. The repository offers tools for image editing, video creation, and text generation, with options for customizing parameters and outputs. Users can access different AI models for specific tasks like image generation, video editing, and text creation.
ai-chatbot
Next.js AI Chatbot is an open-source app template for building AI chatbots using Next.js, Vercel AI SDK, OpenAI, and Vercel KV. It includes features like Next.js App Router, React Server Components, Vercel AI SDK for streaming chat UI, support for various AI models, Tailwind CSS styling, Radix UI for headless components, chat history management, rate limiting, session storage with Vercel KV, and authentication with NextAuth.js. The template allows easy deployment to Vercel and customization of AI model providers.
web-llm-chat
WebLLM Chat is a private AI chat interface that combines WebLLM with a user-friendly design, leveraging WebGPU to run large language models natively in your browser. It offers browser-native AI experience with WebGPU acceleration, guaranteed privacy as all data processing happens locally, offline accessibility, user-friendly interface with markdown support, and open-source customization. The project aims to democratize AI technology by making powerful tools accessible directly to end-users, enhancing the chatting experience and broadening the scope for deployment of self-hosted and customizable language models.
echosharp
EchoSharp is an open-source library designed for near-real-time audio processing, orchestrating different AI models seamlessly for various audio analysis scopes. It focuses on flexibility and performance, allowing near-real-time Transcription and Translation by integrating components for Speech-to-Text and Voice Activity Detection. With interchangeable components, easy orchestration, and first-party components like Whisper.net, SileroVad, OpenAI Whisper, AzureAI SpeechServices, WebRtcVadSharp, Onnx.Whisper, and Onnx.Sherpa, EchoSharp provides efficient audio analysis solutions for developers.
Open-WebUI-Functions
Open-WebUI-Functions is a collection of Python-based functions that extend Open WebUI with custom pipelines, filters, and integrations. Users can interact with AI models, process data efficiently, and customize the Open WebUI experience. It includes features like custom pipelines, data processing filters, Azure AI support, N8N workflow integration, flexible configuration, secure API key management, and support for both streaming and non-streaming processing. The functions require an active Open WebUI instance, may need external AI services like Azure AI, and admin access for installation. Security features include automatic encryption of sensitive information like API keys. Pipelines include Azure AI Foundry, N8N, Infomaniak, and Google Gemini. Filters like Time Token Tracker measure response time and token usage. Integrations with Azure AI, N8N, Infomaniak, and Google are supported. Contributions are welcome, and the project is licensed under Apache License 2.0.
wunjo.wladradchenko.ru
Wunjo AI is a comprehensive tool that empowers users to explore the realm of speech synthesis, deepfake animations, video-to-video transformations, and more. Its user-friendly interface and privacy-first approach make it accessible to both beginners and professionals alike. With Wunjo AI, you can effortlessly convert text into human-like speech, clone voices from audio files, create multi-dialogues with distinct voice profiles, and perform real-time speech recognition. Additionally, you can animate faces using just one photo combined with audio, swap faces in videos, GIFs, and photos, and even remove unwanted objects or enhance the quality of your deepfakes using the AI Retouch Tool. Wunjo AI is an all-in-one solution for your voice and visual AI needs, offering endless possibilities for creativity and expression.
Open-Sora-Plan
Open-Sora-Plan is a project that aims to create a simple and scalable repo to reproduce Sora (OpenAI, but we prefer to call it "ClosedAI"). The project is still in its early stages, but the team is working hard to improve it and make it more accessible to the open-source community. The project is currently focused on training an unconditional model on a landscape dataset, but the team plans to expand the scope of the project in the future to include text2video experiments, training on video2text datasets, and controlling the model with more conditions.
Rewind-AI-Main
Rewind AI is a free and open-source AI-powered video editing tool that allows users to easily create and edit videos. It features a user-friendly interface, a wide range of editing tools, and support for a variety of video formats. Rewind AI is perfect for beginners and experienced video editors alike.
For similar jobs
LLMStack
LLMStack is a no-code platform for building generative AI agents, workflows, and chatbots. It allows users to connect their own data, internal tools, and GPT-powered models without any coding experience. LLMStack can be deployed to the cloud or on-premise and can be accessed via HTTP API or triggered from Slack or Discord.
daily-poetry-image
Daily Chinese ancient poetry and AI-generated images powered by Bing DALL-E-3. GitHub Action triggers the process automatically. Poetry is provided by Today's Poem API. The website is built with Astro.
exif-photo-blog
EXIF Photo Blog is a full-stack photo blog application built with Next.js, Vercel, and Postgres. It features built-in authentication, photo upload with EXIF extraction, photo organization by tag, infinite scroll, light/dark mode, automatic OG image generation, a CMD-K menu with photo search, experimental support for AI-generated descriptions, and support for Fujifilm simulations. The application is easy to deploy to Vercel with just a few clicks and can be customized with a variety of environment variables.
SillyTavern
SillyTavern is a user interface you can install on your computer (and Android phones) that allows you to interact with text generation AIs and chat/roleplay with characters you or the community create. SillyTavern is a fork of TavernAI 1.2.8 which is under more active development and has added many major features. At this point, they can be thought of as completely independent programs.
Twitter-Insight-LLM
This project enables you to fetch liked tweets from Twitter (using Selenium), save it to JSON and Excel files, and perform initial data analysis and image captions. This is part of the initial steps for a larger personal project involving Large Language Models (LLMs).
AISuperDomain
Aila Desktop Application is a powerful tool that integrates multiple leading AI models into a single desktop application. It allows users to interact with various AI models simultaneously, providing diverse responses and insights to their inquiries. With its user-friendly interface and customizable features, Aila empowers users to engage with AI seamlessly and efficiently. Whether you're a researcher, student, or professional, Aila can enhance your AI interactions and streamline your workflow.
ChatGPT-On-CS
This project is an intelligent dialogue customer service tool based on a large model, which supports access to platforms such as WeChat, Qianniu, Bilibili, Douyin Enterprise, Douyin, Doudian, Weibo chat, Xiaohongshu professional account operation, Xiaohongshu, Zhihu, etc. You can choose GPT3.5/GPT4.0/ Lazy Treasure Box (more platforms will be supported in the future), which can process text, voice and pictures, and access external resources such as operating systems and the Internet through plug-ins, and support enterprise AI applications customized based on their own knowledge base.
obs-localvocal
LocalVocal is a live-streaming AI assistant plugin for OBS that allows you to transcribe audio speech into text and perform various language processing functions on the text using AI / LLMs (Large Language Models). It's privacy-first, with all data staying on your machine, and requires no GPU, cloud costs, network, or downtime.







