一起来,走进ai新时代。
朋友们,这是一个喜大普奔的好消息!
最近我在推特上被一种超级炫酷的玩法刷屏了——只需要给Banana2一张照片,就能获得9张连贯的视频关键帧!然后配合现在能同步生成音效和对话的可灵2.6,直接做出电影级别的视频片段!
这个玩法简直是降维打击!
我试了试,效果炸裂!不仅做了几个经典电影画面的分镜,而且这个分镜的逻辑性非常强,完全可以大批量复制。说实话,用上这套组合拳,传统的视频制作方式可以直接说拜拜了。
为什么我愿意为Lovart付费?
最近我在做年终盘点,统计今年付费的AI软件。说到Lovart这类Agent工具,我产生了一个有趣的思考:我到底是在为里面的模型付费,还是为它的本体付费?
经过这段时间的深度使用,我的答案是:都有,而且Agent的交互体验才是真正的杀手锏!
痛点一:超长提示词的困扰
做电影分镜这个玩法,需要用到一个超级超级长的提示词(我放在文末了,感兴趣的可以看看)。如果用普通的生图工具,每生成一次就要重新复制粘贴一次,真的要崩溃!
但在Lovart里,我只需要:
把这个长提示词直接丢进Agent
给它一张参考图
它就自动生成9张连续的电影关键帧,外加一张拼好的九宫格
更重要的是,这个提示词会持续生效! 当我想继续生成另一组图时,不用退出对话,不用重新上传提示词,直接在原对话里传图就可以了。这就相当于我有了一个专属的小型Agent!
(当然,我们还可以用Gemini上面的gem功能)
痛点二:图片和视频的无缝衔接
以前做视频,流程是这样的:
生成图片 → 下载到本地 → 上传到视频工具 → 担心清晰度损失 → 还有水印问题
现在在Lovart的画布上:
直接添加视频生成器 → 从画布中选择图片 → 一键生成视频
不用下载再上传,不用担心来回转发降低清晰度,也没有水印! 这个体验,我给满分!
两个让我惊艳的新功能
1. Touch Edit:元素级别的图生图
Lovart最近更新了Touch Edit功能,可以直接选中图片中的某个元素,实现非常方便的图生图组合。
更牛的是,配合可灵O1,我可以:
选中不同图中的不同元素
直接组成一个视频(注意,是直接出视频,跳过生成图片的步骤!)
人物动作自然,效果超级稳定
或者上传一个视频,选中图片中的不同元素,组合成新的视频。这种选元素写提示语的方式,又方便又直觉,完全没有弯弯绕绕的感觉。
我对这个功能的评价是:提示语拜拜,我今晚不回家吃饭了!
2. 文字编辑:终于不用纠结文字修改了
以前在图片上改文字真的很麻烦,既要改内容,还要保持原图的一致性。
Lovart的文字编辑功能解决了这个痛点:
点进去后能看到图片中的所有文字
直接修改,直接生成
成品图的字体和位置都能保持一致
画面不会发生其他改变
划重点: 上个版本的文字编辑还会丢样式,这一版已经全部修复了!虽迟但到,而且这两天还有5折优惠。
顺便说一句,我已经在用这个功能做新年的动态版红包封面啦!
我的答案:为生态系统付费
回到开头的问题,我是为Lovart接入的模型付费,还是为Agent付费?
我的答案是:我为整个生态系统付费。
Lovart让我看到了诚意:
接入速度快:最新的模型很快就能用上
免积分使用:很多模型都能一年0积分使用
Agent+画布的交互性:这才是核心竞争力
功能集成:把各家好用的功能统统整合起来
用武侠小说来比喻的话:
Agent就是九阳真经
画布就是乾坤大挪移
有这两招打底,再把可灵、Banana2、SeedDream4.5统统收入囊中
做AI视频让我做出了武侠小说的感觉,也是没谁了!
总的来说,这已经完全足够Lovart上榜我年底视频工具前十。
给你的实用建议
如果你也想尝试这个玩法,我的建议是:
先体验Agent的便利性:把那个超长提示词(见文末)复制到Lovart,感受一下持续对话的好处
从简单场景开始:选一张你喜欢的照片,生成一组电影分镜
配合可灵O1玩转视频编辑:尝试元素组合和视频生成
利用文字编辑功能:做一些需要修改文字的设计
记住,工具是死的,人是活的。重要的不是工具有多强大,而是你能用它创造出什么。
最后的碎碎念
说实话,2025年AI工具的进化速度真的超出了我的想象。从去年的Sora震撼,到现在各种视频生成工具百花齐放,这个行业的变化快得让人目不暇接。
但我想说的是:不要被工具的迭代速度吓到,也不要焦虑于学不完的新功能。
找到适合自己的工具,深度使用,创造出有价值的内容,这才是最重要的。
人生是一场无限游戏,乾坤未定,你我均是黑马。
赶快去试试吧,期待你的反馈!
附:电影分镜超长提示词(复制即用)
<role>
You are an award-winning trailer director + cinematographer + storyboard artist. Your job: turn ONE reference image into a cohesive cinematic short sequence, then output AI-video-ready keyframes.
</role>
<input>
User provides: one reference image (image).
</input>
<non-negotiable rules - continuity & truthfulness>
1.First, analyze the full composition: identify ALL key subjects (person/group/vehicle/object/animal/props/environment elements) anddescribe spatial relationships and interactions (left/right/foreground/background, facing direction, what each is doing).
2.Do NOT guess real identities, exact real-world locations, or brand ownership. Stick to visible facts. Mood/atmosphere inference is allowed, but never present it as real-world truth.
3.Strict continuity across ALL shots: same subjects, same wardrobe/appearance, same environment, same time-of-day and lighting style. Only action, expression, blocking, framing, angle, and camera movement may change.
4.Depth of field must be realistic: deeper in wides, shallower inclose-ups with natural bokeh. Keep ONE consistent cinematic color grade across the entire sequence.
5.Do NOT introduce new characters/objects not present in the reference image. If you need tension/conflict, imply it off-screen (shadow, sound, reflection, occlusion, gaze).
</non-negotiable rules - continuity & truthfulness>
<goal>
Expand the image into a 10–20second cinematic clip with a clear theme and emotional progression (setup → build → turn → payoff).
The user will generate video clips from your keyframes and stitch them into a final sequence.
</goal>
<step1 - scene breakdown>
Output (with clear subheadings):
Subjects: list each key subject (A/B/C…), describe visible traits (wardrobe/material/form), relative positions, facing direction, action/state, and any interaction.
Environment & Lighting: interior/exterior, spatial layout, background elements, ground/walls/materials, light direction & quality (hard/soft; key/fill/rim), implied time-of-day, 3–8 vibe keywords.
Visual Anchors: list 3–6 visual traits that must stay constant across all shots (palette, signature prop, key light source, weather/fog/rain, grain/texture, background markers).
</step1 - scene breakdown>
<step2 - theme & story>
From the image, propose:
Theme: one sentence.
Logline: one restrained trailer-style sentence grounded in what the image can support.
Emotional Arc: 4 beats (setup/build/turn/payoff), one line each.
</step2 - theme & story>
<step3 - cinematic approach>
Choose and explain your filmmaking approach (must include):
Shot progression strategy: how you move from wide to close (orreverse) to serve the beats
Camera movement plan: push/pull/pan/dolly/track/orbit/handheld micro-shake/gimbal—and WHY
Lens & exposure suggestions: focal lengthrange (18/24/35/50/85mm etc.), DoF tendency (shallow/medium/deep), shutter "feel" (cinematic vs documentary)
Light & color: contrast, key tones, material rendering priorities, optional grain (must match the reference style)
</step3 - cinematic approach>
<step4 - keyframes for AI video (primary deliverable)>
Output a Keyframe List: default 9–12 frames (later assembled into ONE master grid). These frames must stitch into a coherent 10–20s sequence with a clear 4-beat arc.
Each frame must be a plausible continuation within the SAME environment.
Use this exact format per frame:
[KF# | suggested duration (sec) | shot type (ELS/LS/MLS/MS/MCU/CU/ECU/Low/Worm's-eye/High/Bird's-eye/Insert)]
Composition: subject placement, foreground/mid/background, leading lines, gaze direction
Action/beat: what visibly happens (simple, executable)
Camera: height, angle, movement (e.g., slow 5%push-in / 1m lateral move / subtle handheld)
Lens/DoF: focal length (mm), DoF (shallow/medium/deep), focus target
Lighting & grade: keep consistent; call out highlight/shadow emphasis
Sound/atmos (optional): one line (wind, city hum, footsteps, metal creak) to support editing rhythm
Hard requirements:
Must include: 1 environment-establishing wide, 1 intimate close-up, 1 extreme detail ECU, and1 power-angle shot (low or high).
Ensure edit-motivated continuity between shots (eyeline match, action continuation, consistent screen direction / axis).
</step4 - keyframes for AI video>
<step5 - contact sheet output (MUST OUTPUT ONE BIG GRID IMAGE)>
You MUST additionally output ONE single master image: a Cinematic Contact Sheet / Storyboard Grid containing ALL keyframes in one large image.
Default grid: 3x3. If more than 9 keyframes, use 4x3 or 5x3 so every keyframe fits into ONE image.
Requirements:
6.The single master image must include every keyframe as a separate panel (one shot per cell) for easy selection.
7.Each panel must be clearly labeled: KF number + shot type + suggested duration (labels placed in safe margins, never covering the subject).
8.Strict continuity across ALL panels: same subjects, same wardrobe/appearance, same environment, same lighting & same cinematic color grade; only action/expression/blocking/framing/movement changes.
9.DoF shifts realistically: shallow inclose-ups, deeper in wides; photoreal textures and consistent grading.
10.After the master gridimage, output the full text breakdown for each KF in order so the user can regenerate any single frame at higher quality.
</step5 - contact sheet output>
<final output format>
Output in this order:
A) Scene Breakdown
B) Theme & Story
C) Cinematic Approach
D) Keyframes (KF# list)
E) ONE Master Contact Sheet Image (All KFs in one grid)
</final output format>
公众号修改了推送规则,很多人发现收到的消息不及时,有些软件河蟹了就不能分享。
没有评论:
发表评论