linyq c065800072 feat(webui): 大改动标记1
-重构音频设置面板,增加语音音量、背景音乐等设置
- 添加背景音乐文件选择功能
- 优化字幕设置,支持自定义字体和样式
-调整视频生成流程,支持新音频设置
- 更新文档示例,反映新功能
2024-12-10 18:33:44 +08:00

443 lines
14 KiB
Python
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

import traceback
import pysrt
from typing import Optional
from typing import List
from loguru import logger
from moviepy.editor import *
from PIL import ImageFont
from contextlib import contextmanager
from moviepy.editor import (
VideoFileClip,
AudioFileClip,
TextClip,
CompositeVideoClip,
CompositeAudioClip
)
from app.models.schema import VideoAspect
def wrap_text(text, max_width, font, fontsize=60):
"""
文本自动换行处理
Args:
text: 待处理的文本
max_width: 最大宽度
font: 字体文件路径
fontsize: 字体大小
Returns:
tuple: (换行后的文本, 文本高度)
"""
# 创建字体对象
font = ImageFont.truetype(font, fontsize)
def get_text_size(inner_text):
inner_text = inner_text.strip()
left, top, right, bottom = font.getbbox(inner_text)
return right - left, bottom - top
width, height = get_text_size(text)
if width <= max_width:
return text, height
logger.debug(f"换行文本, 最大宽度: {max_width}, 文本宽度: {width}, 文本: {text}")
processed = True
_wrapped_lines_ = []
words = text.split(" ")
_txt_ = ""
for word in words:
_before = _txt_
_txt_ += f"{word} "
_width, _height = get_text_size(_txt_)
if _width <= max_width:
continue
else:
if _txt_.strip() == word.strip():
processed = False
break
_wrapped_lines_.append(_before)
_txt_ = f"{word} "
_wrapped_lines_.append(_txt_)
if processed:
_wrapped_lines_ = [line.strip() for line in _wrapped_lines_]
result = "\n".join(_wrapped_lines_).strip()
height = len(_wrapped_lines_) * height
# logger.warning(f"wrapped text: {result}")
return result, height
_wrapped_lines_ = []
chars = list(text)
_txt_ = ""
for word in chars:
_txt_ += word
_width, _height = get_text_size(_txt_)
if _width <= max_width:
continue
else:
_wrapped_lines_.append(_txt_)
_txt_ = ""
_wrapped_lines_.append(_txt_)
result = "\n".join(_wrapped_lines_).strip()
height = len(_wrapped_lines_) * height
logger.debug(f"换行文本: {result}")
return result, height
@contextmanager
def manage_clip(clip):
"""
视频片段资源管理器
Args:
clip: 视频片段对象
Yields:
VideoFileClip: 视频片段对象
"""
try:
yield clip
finally:
clip.close()
del clip
def combine_clip_videos(combined_video_path: str,
video_paths: List[str],
video_ost_list: List[int],
list_script: list,
video_aspect: VideoAspect = VideoAspect.portrait,
threads: int = 2,
) -> str:
"""
合并子视频
Args:
combined_video_path: 合并后的存储路径
video_paths: 子视频路径列表
video_ost_list: 原声播放列表 (0: 不保留原声, 1: 只保留原声, 2: 保留原声并保留解说)
list_script: 剪辑脚本
video_aspect: 屏幕比例
threads: 线程数
Returns:
str: 合并后的视频路径
"""
from app.utils.utils import calculate_total_duration
audio_duration = calculate_total_duration(list_script)
logger.info(f"音频的最大持续时间: {audio_duration} s")
output_dir = os.path.dirname(combined_video_path)
aspect = VideoAspect(video_aspect)
video_width, video_height = aspect.to_resolution()
clips = []
for video_path, video_ost in zip(video_paths, video_ost_list):
try:
clip = VideoFileClip(video_path)
if video_ost == 0: # 不保留原声
clip = clip.without_audio()
# video_ost 为 1 或 2 时都保留原声,不需要特殊处理
clip = clip.set_fps(30)
# 处理视频尺寸
clip_w, clip_h = clip.size
if clip_w != video_width or clip_h != video_height:
clip = resize_video_with_padding(
clip,
target_width=video_width,
target_height=video_height
)
logger.info(f"视频 {video_path} 已调整尺寸为 {video_width} x {video_height}")
clips.append(clip)
except Exception as e:
logger.error(f"处理视频 {video_path} 时出错: {str(e)}")
continue
if not clips:
raise ValueError("没有有效的视频片段可以合并")
try:
video_clip = concatenate_videoclips(clips)
video_clip = video_clip.set_fps(30)
logger.info("开始合并视频... (过程中出现 UserWarning: 不必理会)")
video_clip.write_videofile(
filename=combined_video_path,
threads=threads,
logger=None,
audio_codec="aac",
fps=30,
temp_audiofile=os.path.join(output_dir, "temp-audio.m4a")
)
finally:
# 确保资源被正确放
video_clip.close()
for clip in clips:
clip.close()
logger.success("视频合并完成")
return combined_video_path
def resize_video_with_padding(clip, target_width: int, target_height: int):
"""
调整视频尺寸并添加黑边
Args:
clip: 视频片段
target_width: 目标宽度
target_height: 目标高度
Returns:
CompositeVideoClip: 调整尺寸后的视频
"""
clip_ratio = clip.w / clip.h
target_ratio = target_width / target_height
if clip_ratio == target_ratio:
return clip.resize((target_width, target_height))
if clip_ratio > target_ratio:
scale_factor = target_width / clip.w
else:
scale_factor = target_height / clip.h
new_width = int(clip.w * scale_factor)
new_height = int(clip.h * scale_factor)
clip_resized = clip.resize(newsize=(new_width, new_height))
background = ColorClip(
size=(target_width, target_height),
color=(0, 0, 0)
).set_duration(clip.duration)
return CompositeVideoClip([
background,
clip_resized.set_position("center")
])
def loop_audio_clip(audio_clip: AudioFileClip, target_duration: float) -> AudioFileClip:
"""
循环音频片段直到达到目标时长
参数:
audio_clip: 原始音频片段
target_duration: 目标时长(秒)
返回:
循环后的音频片段
"""
# 计算需要循环的次数
loops_needed = int(target_duration / audio_clip.duration) + 1
# 创建足够长的音频
extended_audio = audio_clip
for _ in range(loops_needed - 1):
extended_audio = CompositeAudioClip([
extended_audio,
audio_clip.set_start(extended_audio.duration)
])
# 裁剪到目标时长
return extended_audio.subclip(0, target_duration)
def generate_video_v3(
video_path: str,
subtitle_path: Optional[str] = None,
bgm_path: Optional[str] = None,
narration_path: Optional[str] = None,
output_path: str = "output.mp4",
# 音量相关参数
volume_config: dict = None,
# 字幕相关参数
subtitle_style: dict = None,
font_path: Optional[str] = None
) -> None:
"""
合并视频素材包括视频、字幕、BGM和解说音频
参数:
video_path: 原视频文件路径
subtitle_path: SRT字幕文件路径可选
bgm_path: 背景音乐文件路径(可选)
narration_path: 解说音频文件路径(可选)
output_path: 输出文件路径
volume_config: 音量配置字典,可包含以下键:
- original: 原声音量0-1默认1.0
- bgm: BGM音量0-1默认0.3
- narration: 解说音量0-1默认1.0
subtitle_style: 字幕样式配置字典,可包含以下键:
- font: 字体名称
- fontsize: 字体大小
- color: 字体颜色
- stroke_color: 描边颜色
- stroke_width: 描边宽度
- bg_color: 背景色
- position: 位置支持 'top'/'center'/'bottom' 或 (x,y) 坐标
- method: 文字渲染方法
font_path: 字体文件路径(.ttf/.otf 等格式)
"""
# 检查视频文件是否存在
if not os.path.exists(video_path):
raise FileNotFoundError(f"视频文件不存在: {video_path}")
# 设置默认音量配置
default_volume = {
'original': 1.0, # 原声音量
'bgm': 0.3, # BGM音量
'narration': 1.0 # 解说音量
}
# 更新音量配置
if volume_config:
default_volume.update(volume_config)
# 加载视频
video = VideoFileClip(video_path)
subtitle_clips = []
# 处理字幕(如果提供)
if subtitle_path:
if os.path.exists(subtitle_path):
# 检查字体文件
if font_path and not os.path.exists(font_path):
logger.info(f"警告:字体文件不存在: {font_path},将使用系统默认字体")
font_path = 'Arial'
# 设置默认字幕样式
default_style = {
'font': font_path if font_path else 'Arial',
'fontsize': 24,
'color': 'white',
'stroke_color': 'black',
'stroke_width': 1,
'bg_color': None,
'position': ('center', 'bottom'),
'method': 'label'
}
if subtitle_style:
if font_path and 'font' not in subtitle_style:
subtitle_style['font'] = font_path
default_style.update(subtitle_style)
try:
subs = pysrt.open(subtitle_path)
logger.info(f"读取到 {len(subs)} 条字幕")
for index, sub in enumerate(subs):
start_time = sub.start.ordinal / 1000
end_time = sub.end.ordinal / 1000
try:
# 检查字幕文本是否为空
if not sub.text or sub.text.strip() == '':
logger.info(f"警告:第 {index + 1} 条字幕内容为空,已跳过")
continue
# 处理字幕文本:确保是字符串,并处理可能的列表情况
if isinstance(sub.text, (list, tuple)):
subtitle_text = ' '.join(str(item) for item in sub.text if item is not None)
else:
subtitle_text = str(sub.text)
subtitle_text = subtitle_text.strip()
if not subtitle_text:
logger.info(f"警告:第 {index + 1} 条字幕处理后为空,已跳过")
continue
# 计算位置
if isinstance(default_style['position'], tuple):
pos_x, pos_y = default_style['position']
if isinstance(pos_y, float):
y_pos = int(video.h * pos_y)
position = (pos_x, y_pos)
else:
position = default_style['position']
else:
position = default_style['position']
# 创建基本的 TextClip
text_clip = (TextClip(
subtitle_text,
font=default_style['font'],
fontsize=default_style['fontsize'],
color=default_style['color']
)
.set_position(position)
.set_duration(end_time - start_time)
.set_start(start_time))
subtitle_clips.append(text_clip)
except Exception as e:
logger.info(f"警告:创建第 {index + 1} 条字幕时出错: {str(e)}")
logger.info(f"成功创建 {len(subtitle_clips)} 条字幕剪辑")
except Exception as e:
logger.info(f"警告:处理字幕文件时出错: {str(e)}")
else:
logger.info(f"提示:字幕文件不存在: {subtitle_path}")
# 合并音频
audio_clips = []
# 添加原声(设置音量)
if video.audio is not None:
original_audio = video.audio.volumex(default_volume['original'])
audio_clips.append(original_audio)
# 添加BGM如果提供
if bgm_path:
bgm = AudioFileClip(bgm_path)
if bgm.duration < video.duration:
bgm = loop_audio_clip(bgm, video.duration)
else:
bgm = bgm.subclip(0, video.duration)
bgm = bgm.volumex(default_volume['bgm'])
audio_clips.append(bgm)
# 添加解说音频(如果提供)
if narration_path:
narration = AudioFileClip(narration_path).volumex(default_volume['narration'])
audio_clips.append(narration)
# 合成最终视频(包含字幕)
if subtitle_clips:
final_video = CompositeVideoClip([video] + subtitle_clips, size=video.size)
else:
logger.info("警告:没有字幕被添加到视频中")
final_video = video
if audio_clips:
final_audio = CompositeAudioClip(audio_clips)
final_video = final_video.set_audio(final_audio)
# 导出视频
logger.info("开始导出视频...") # 调试信息
final_video.write_videofile(
output_path,
codec='libx264',
audio_codec='aac',
fps=video.fps
)
logger.info(f"视频已导出到: {output_path}") # 调试信息
# 清理资源
video.close()
for clip in subtitle_clips:
clip.close()
if bgm_path:
bgm.close()
if narration_path:
narration.close()