| | import imageio, librosa |
| | import torch |
| | from PIL import Image |
| | from tqdm import tqdm |
| | import numpy as np |
| |
|
| |
|
| | def resize_image_by_longest_edge(image_path, target_size): |
| | image = Image.open(image_path).convert("RGB") |
| | width, height = image.size |
| | scale = target_size / max(width, height) |
| | new_size = (int(width * scale), int(height * scale)) |
| | return image.resize(new_size, Image.LANCZOS) |
| |
|
| |
|
| | def save_video(frames, save_path, fps, quality=9, ffmpeg_params=None): |
| | writer = imageio.get_writer( |
| | save_path, fps=fps, quality=quality, ffmpeg_params=ffmpeg_params |
| | ) |
| | for frame in tqdm(frames, desc="Saving video"): |
| | frame = np.array(frame) |
| | writer.append_data(frame) |
| | writer.close() |
| |
|
| |
|
| | def get_audio_features(wav2vec, audio_processor, audio_path, fps, num_frames): |
| | sr = 16000 |
| | audio_input, sample_rate = librosa.load(audio_path, sr=sr) |
| |
|
| | start_time = 0 |
| | |
| | end_time = num_frames / fps |
| |
|
| | start_sample = int(start_time * sr) |
| | end_sample = int(end_time * sr) |
| |
|
| | try: |
| | audio_segment = audio_input[start_sample:end_sample] |
| | except: |
| | audio_segment = audio_input |
| |
|
| | input_values = audio_processor( |
| | audio_segment, sampling_rate=sample_rate, return_tensors="pt" |
| | ).input_values.to("cuda") |
| |
|
| | with torch.no_grad(): |
| | fea = wav2vec(input_values).last_hidden_state |
| |
|
| | return fea |
| |
|