format all code
This commit is contained in:
parent
8b97fa5ebd
commit
937c938837
111
main.py
111
main.py
@ -24,14 +24,10 @@ from src.math.distribution import create_distribution
|
||||
|
||||
log = structlog.get_logger()
|
||||
|
||||
EDITORS = {
|
||||
'amplitude': AmplitudeEditor,
|
||||
'sentiment': SentimentEditor
|
||||
}
|
||||
EDITORS = {"amplitude": AmplitudeEditor, "sentiment": SentimentEditor}
|
||||
|
||||
ERROR_FUNCS = {"quadratic": quadratic_loss}
|
||||
|
||||
ERROR_FUNCS = {
|
||||
'quadratic': quadratic_loss
|
||||
}
|
||||
|
||||
def main(args):
|
||||
# Check video existance
|
||||
@ -46,7 +42,9 @@ def main(args):
|
||||
if intro_file is not None:
|
||||
intro_vid_path = Path(intro_file)
|
||||
if not in_vid_path.is_file():
|
||||
log.error("the specified input path does not exist for the intro", path=intro_file)
|
||||
log.error(
|
||||
"the specified input path does not exist for the intro", path=intro_file
|
||||
)
|
||||
sys.exit(-1)
|
||||
log.info("found intro", input_video=intro_file)
|
||||
|
||||
@ -54,7 +52,9 @@ def main(args):
|
||||
if outro_file is not None:
|
||||
outro_vid_path = Path(outro_file)
|
||||
if not outro_vid_path.is_file():
|
||||
log.error("the specified input path does not exist for the outro", path=outro_file)
|
||||
log.error(
|
||||
"the specified input path does not exist for the outro", path=outro_file
|
||||
)
|
||||
sys.exit(-1)
|
||||
log.info("found outro", input_video=outro_file)
|
||||
|
||||
@ -62,7 +62,7 @@ def main(args):
|
||||
# and as a simple way to generate temp file names
|
||||
sha1 = hashlib.sha1()
|
||||
BUF_SIZE = 1655360
|
||||
with open(in_vid_path, 'rb') as f:
|
||||
with open(in_vid_path, "rb") as f:
|
||||
while True:
|
||||
data = f.read(BUF_SIZE)
|
||||
if not data:
|
||||
@ -73,7 +73,9 @@ def main(args):
|
||||
temp_file_name = f"ale-{temp_file_name}"
|
||||
|
||||
# Prepare the input video
|
||||
audio_path, audio_cached = extract_audio_from_video(str(in_vid_path.resolve()), temp_file_name)
|
||||
audio_path, audio_cached = extract_audio_from_video(
|
||||
str(in_vid_path.resolve()), temp_file_name
|
||||
)
|
||||
if audio_cached:
|
||||
log.info("using cached audio file", cache_path=audio_path)
|
||||
else:
|
||||
@ -111,9 +113,13 @@ def main(args):
|
||||
complete = False
|
||||
iterations = 0
|
||||
while not complete:
|
||||
large_distribution = create_distribution(large_window_center, spread_multiplier, parallelism)
|
||||
large_distribution = create_distribution(
|
||||
large_window_center, spread_multiplier, parallelism
|
||||
)
|
||||
np.random.shuffle(large_distribution)
|
||||
small_distribution = create_distribution(small_window_center, spread_multiplier, parallelism)
|
||||
small_distribution = create_distribution(
|
||||
small_window_center, spread_multiplier, parallelism
|
||||
)
|
||||
np.random.shuffle(small_distribution)
|
||||
|
||||
# Fire off workers to generate edits
|
||||
@ -127,7 +133,7 @@ def main(args):
|
||||
editor.edit,
|
||||
pair[0] if pair[0] > pair[1] else pair[1],
|
||||
pair[1] if pair[0] > pair[1] else pair[0],
|
||||
vars(args)
|
||||
vars(args),
|
||||
)
|
||||
)
|
||||
for future in concurrent.futures.as_completed(futures):
|
||||
@ -149,18 +155,43 @@ def main(args):
|
||||
index_min = min(range(len(costs)), key=costs.__getitem__)
|
||||
large_window_center = moment_results[index_min][1]
|
||||
small_window_center = moment_results[index_min][2]
|
||||
log.info("batch complete", best_large=large_window_center, best_small=small_window_center, duration=durations[index_min])
|
||||
if durations[index_min] > desired * 0.95 and desired * 1.05 > durations[index_min]:
|
||||
log.info("found edit within target duration", target=desired, duration=durations[index_min])
|
||||
log.info(
|
||||
"batch complete",
|
||||
best_large=large_window_center,
|
||||
best_small=small_window_center,
|
||||
duration=durations[index_min],
|
||||
)
|
||||
if (
|
||||
durations[index_min] > desired * 0.95
|
||||
and desired * 1.05 > durations[index_min]
|
||||
):
|
||||
log.info(
|
||||
"found edit within target duration",
|
||||
target=desired,
|
||||
duration=durations[index_min],
|
||||
)
|
||||
out_path = Path(args.destination)
|
||||
log.info("rendering...")
|
||||
start = time.time()
|
||||
render_moments(moment_results[index_min][0], str(in_vid_path.resolve()), str(out_path.resolve()), intro_path=intro_file, parallelism=args.parallelism)
|
||||
log.info("render complete", duration=time.time() - start, output=str(out_path.resolve()))
|
||||
render_moments(
|
||||
moment_results[index_min][0],
|
||||
str(in_vid_path.resolve()),
|
||||
str(out_path.resolve()),
|
||||
intro_path=intro_file,
|
||||
parallelism=args.parallelism,
|
||||
)
|
||||
log.info(
|
||||
"render complete",
|
||||
duration=time.time() - start,
|
||||
output=str(out_path.resolve()),
|
||||
)
|
||||
sys.exit(0)
|
||||
iterations = iterations + parallelism
|
||||
if iterations > 50000:
|
||||
log.error("could not find a viable edit in the target duration, try other params", target=desired)
|
||||
log.error(
|
||||
"could not find a viable edit in the target duration, try other params",
|
||||
target=desired,
|
||||
)
|
||||
sys.exit(-4)
|
||||
spread_multiplier = spread_multiplier - spread_decay
|
||||
if spread_multiplier < 0:
|
||||
@ -173,15 +204,23 @@ def main(args):
|
||||
|
||||
if __name__ == "__main__":
|
||||
parser = argparse.ArgumentParser(
|
||||
prog="ALE", description="ALE: Automatic Linear Editor.",
|
||||
formatter_class=partial(argparse.HelpFormatter, width=100)
|
||||
prog="ALE",
|
||||
description="ALE: Automatic Linear Editor.",
|
||||
formatter_class=partial(argparse.HelpFormatter, width=100),
|
||||
)
|
||||
parser.add_argument("file", help="Path to the video file to edit")
|
||||
parser.add_argument(
|
||||
"duration", help="Target length of the edit, in seconds", type=int
|
||||
)
|
||||
parser.add_argument("destination", help="Edited video save location")
|
||||
subparsers = parser.add_subparsers(
|
||||
dest="editor", help="The editing algorithm to use"
|
||||
)
|
||||
parser.add_argument('file', help='Path to the video file to edit')
|
||||
parser.add_argument('duration', help='Target length of the edit, in seconds', type=int)
|
||||
parser.add_argument('destination', help='Edited video save location')
|
||||
subparsers = parser.add_subparsers(dest='editor', help='The editing algorithm to use')
|
||||
|
||||
parser_audio_amp = subparsers.add_parser('amplitude', help='The amplitude editor uses audio amplitude moving averages to find swings from relatively quiet moments to loud moments. This is useful in videos where long moments of quiet are interspersed with loud action filled moments.')
|
||||
parser_audio_amp = subparsers.add_parser(
|
||||
"amplitude",
|
||||
help="The amplitude editor uses audio amplitude moving averages to find swings from relatively quiet moments to loud moments. This is useful in videos where long moments of quiet are interspersed with loud action filled moments.",
|
||||
)
|
||||
parser_audio_amp.add_argument(
|
||||
"--factor",
|
||||
default=16000,
|
||||
@ -190,7 +229,10 @@ if __name__ == "__main__":
|
||||
type=int,
|
||||
)
|
||||
|
||||
parser_sentiment = subparsers.add_parser('sentiment', help='The sentiment editor transcribes the speech in a video and runs sentiment analysis on the resulting text. Using moving averages, large swings in sentiment can be correlated to controversial or exciting moments. A GPU with CUDA is recommended for fast results.')
|
||||
parser_sentiment = subparsers.add_parser(
|
||||
"sentiment",
|
||||
help="The sentiment editor transcribes the speech in a video and runs sentiment analysis on the resulting text. Using moving averages, large swings in sentiment can be correlated to controversial or exciting moments. A GPU with CUDA is recommended for fast results.",
|
||||
)
|
||||
parser_sentiment.add_argument(
|
||||
"--model",
|
||||
default="base",
|
||||
@ -199,8 +241,17 @@ if __name__ == "__main__":
|
||||
choices=["base", "tiny", "small", "medium", "large"],
|
||||
)
|
||||
|
||||
parser.add_argument("-p", "--parallelism", dest="parallelism", type=int, default=multiprocessing.cpu_count() - 2, help="The number of cores to use, defaults to N - 2 cores.")
|
||||
parser.add_argument("--cost-function", dest="cost", choices=ERROR_FUNCS.keys(), default='quadratic')
|
||||
parser.add_argument(
|
||||
"-p",
|
||||
"--parallelism",
|
||||
dest="parallelism",
|
||||
type=int,
|
||||
default=multiprocessing.cpu_count() - 2,
|
||||
help="The number of cores to use, defaults to N - 2 cores.",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--cost-function", dest="cost", choices=ERROR_FUNCS.keys(), default="quadratic"
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
"--min-duration",
|
||||
|
@ -4,6 +4,7 @@ from ..common import find_moving_average_highlights
|
||||
import numpy as np
|
||||
import structlog
|
||||
|
||||
|
||||
class AmplitudeEditor:
|
||||
def __init__(self, video_path, audio_path, params):
|
||||
self.logger = structlog.get_logger("amplitude")
|
||||
@ -21,8 +22,13 @@ class AmplitudeEditor:
|
||||
|
||||
def edit(self, large_window, small_window, params):
|
||||
window_factor = self.bitrate / self.factor
|
||||
long_ma = np_moving_average(self.squared_subsample, large_window * window_factor)
|
||||
short_ma = np_moving_average(self.squared_subsample, small_window * window_factor)
|
||||
highlights = find_moving_average_highlights(short_ma, long_ma, self.factor / self.bitrate)
|
||||
long_ma = np_moving_average(
|
||||
self.squared_subsample, large_window * window_factor
|
||||
)
|
||||
short_ma = np_moving_average(
|
||||
self.squared_subsample, small_window * window_factor
|
||||
)
|
||||
highlights = find_moving_average_highlights(
|
||||
short_ma, long_ma, self.factor / self.bitrate
|
||||
)
|
||||
return highlights, large_window, small_window
|
||||
|
||||
|
@ -1,5 +1,6 @@
|
||||
from ..models.moment import Moment
|
||||
|
||||
|
||||
def find_moving_average_highlights(short_ma, long_ma, scaling_factor=1):
|
||||
in_a_clip = False
|
||||
m = None
|
||||
|
@ -12,6 +12,7 @@ from flair.data import Sentence
|
||||
from ...math.average import np_moving_average
|
||||
from ..common import find_moving_average_highlights
|
||||
|
||||
|
||||
@dataclass
|
||||
class TextGlob:
|
||||
start: float
|
||||
@ -19,6 +20,7 @@ class TextGlob:
|
||||
text: str
|
||||
sentiment: float
|
||||
|
||||
|
||||
class SentimentEditor:
|
||||
def __init__(self, video_path, audio_path, params):
|
||||
self.logger = structlog.get_logger("sentiment")
|
||||
@ -30,20 +32,24 @@ class SentimentEditor:
|
||||
self.logger.info("transcribing audio", path=audio_path)
|
||||
self.result = self.model.transcribe(audio_path)
|
||||
|
||||
with open(dest_location, 'w') as fp:
|
||||
with open(dest_location, "w") as fp:
|
||||
json.dump(self.result, fp)
|
||||
else:
|
||||
self.logger.info("cached transcription found", path=dest_location)
|
||||
|
||||
with open(dest_location, 'r') as f:
|
||||
with open(dest_location, "r") as f:
|
||||
self.result = json.load(f)
|
||||
|
||||
self.segments = []
|
||||
for segment in self.result['segments']:
|
||||
self.segments.append(TextGlob(segment['start'], segment['end'], segment['text'], 0))
|
||||
classifier = TextClassifier.load('en-sentiment')
|
||||
for segment in self.result["segments"]:
|
||||
self.segments.append(
|
||||
TextGlob(segment["start"], segment["end"], segment["text"], 0)
|
||||
)
|
||||
classifier = TextClassifier.load("en-sentiment")
|
||||
self.sentiments = []
|
||||
self.logger.info("calculating sentiment on segments", segments=len(self.segments))
|
||||
self.logger.info(
|
||||
"calculating sentiment on segments", segments=len(self.segments)
|
||||
)
|
||||
for segment in self.segments:
|
||||
sentence = Sentence(segment.text)
|
||||
classifier.predict(sentence)
|
||||
@ -59,5 +65,7 @@ class SentimentEditor:
|
||||
window_factor = len(self.sentiments) / end_time
|
||||
long_ma = np_moving_average(self.sentiments, large_window)
|
||||
short_ma = np_moving_average(self.sentiments, small_window)
|
||||
highlights = find_moving_average_highlights(short_ma, long_ma, 1.0 / window_factor)
|
||||
highlights = find_moving_average_highlights(
|
||||
short_ma, long_ma, 1.0 / window_factor
|
||||
)
|
||||
return highlights, large_window, small_window
|
||||
|
@ -1,4 +1,5 @@
|
||||
import numpy as np
|
||||
|
||||
|
||||
def np_moving_average(data: int, window: int) -> np.ndarray:
|
||||
return np.convolve(data, np.ones(int(window)), "valid") / window
|
||||
|
@ -3,5 +3,6 @@ Functions in this file should always target for 0 to be the
|
||||
lowest possible error -> smaller values win
|
||||
"""
|
||||
|
||||
|
||||
def quadratic_loss(target, result):
|
||||
return (target - result) ** 2.0
|
||||
|
@ -1,5 +1,6 @@
|
||||
import numpy as np
|
||||
|
||||
|
||||
def create_distribution(center, spread, count):
|
||||
high = center * (1.0 + spread)
|
||||
low = center - (center * spread)
|
||||
|
@ -4,10 +4,8 @@ from pathlib import Path
|
||||
import numpy as np
|
||||
import scipy.io.wavfile as wav
|
||||
|
||||
def extract_audio_from_video(
|
||||
video_path: str,
|
||||
filename: str
|
||||
):
|
||||
|
||||
def extract_audio_from_video(video_path: str, filename: str):
|
||||
tempdir = tempfile.gettempdir()
|
||||
|
||||
dest_location = f"{tempdir}/{filename}.wav"
|
||||
@ -19,6 +17,7 @@ def extract_audio_from_video(
|
||||
vid.close()
|
||||
return dest_location, False
|
||||
|
||||
|
||||
def process_audio(source_audio_path):
|
||||
rate, data_raw = wav.read(source_audio_path)
|
||||
data_raw = data_raw.astype(np.int32)
|
||||
@ -26,5 +25,6 @@ def process_audio(source_audio_path):
|
||||
duration = len(mono) / rate
|
||||
return mono, duration, rate
|
||||
|
||||
|
||||
def resample(data: np.ndarray, factor: int) -> np.ndarray:
|
||||
return data[::factor].copy()
|
||||
|
@ -9,16 +9,27 @@ def get_subclips(source_video_path, moments):
|
||||
return clips, vid
|
||||
|
||||
|
||||
def render_moments(moments, input_video_path, output_path, intro_path=None, outro_path=None, parallelism=1):
|
||||
def render_moments(
|
||||
moments,
|
||||
input_video_path,
|
||||
output_path,
|
||||
intro_path=None,
|
||||
outro_path=None,
|
||||
parallelism=1,
|
||||
):
|
||||
clips, _ = get_subclips(input_video_path, moments)
|
||||
if intro_path is not None:
|
||||
size = clips[0].size
|
||||
iclip = mp.VideoFileClip(intro_path)
|
||||
iclip.resize(height=size[1])
|
||||
clips.insert(0, iclip)
|
||||
composite = mp.concatenate_videoclips(clips, method='compose')
|
||||
composite = mp.concatenate_videoclips(clips, method="compose")
|
||||
composite.write_videofile(output_path, logger=None, threads=parallelism)
|
||||
|
||||
|
||||
def filter_moments(moments, min_length, max_length):
|
||||
return [m for m in moments if m.get_duration() > min_length and m.get_duration() < max_length]
|
||||
return [
|
||||
m
|
||||
for m in moments
|
||||
if m.get_duration() > min_length and m.get_duration() < max_length
|
||||
]
|
||||
|
@ -1,6 +1,4 @@
|
||||
|
||||
class Moment:
|
||||
|
||||
def __init__(self, start, stop):
|
||||
self.start = start
|
||||
self.stop = stop
|
||||
|
@ -9,12 +9,18 @@ log = structlog.get_logger()
|
||||
|
||||
|
||||
def install_ffmpeg():
|
||||
if not click.confirm('Do you want to install ffmpeg? It is required for ALE.', default=False):
|
||||
log.warn("ffmpeg not installed. Please install it manually or restart ALE. Exiting...")
|
||||
if not click.confirm(
|
||||
"Do you want to install ffmpeg? It is required for ALE.", default=False
|
||||
):
|
||||
log.warn(
|
||||
"ffmpeg not installed. Please install it manually or restart ALE. Exiting..."
|
||||
)
|
||||
sys.exit(0)
|
||||
system = platform.system().lower()
|
||||
if system == "linux":
|
||||
package_manager = "apt-get" if os.path.exists("/etc/apt/sources.list") else "yum"
|
||||
package_manager = (
|
||||
"apt-get" if os.path.exists("/etc/apt/sources.list") else "yum"
|
||||
)
|
||||
command = f"sudo {package_manager} install -y ffmpeg"
|
||||
elif system == "darwin":
|
||||
command = "brew install ffmpeg"
|
||||
|
Loading…
Reference in New Issue
Block a user