From cf5c58e69c56a5c309e009260e1a37ed650e5db4 Mon Sep 17 00:00:00 2001 From: Test User Date: Thu, 19 Feb 2026 02:45:16 +0000 Subject: [PATCH] fix(inference): correct output path in VAE encode success message MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit The print statement incorrectly used the encoded_output tensor variable instead of args.output_path when displaying the save location. Changed: {encoded_output}/encoded.pt → {args.output_path}/encoded.pt --- inference/cli_vae_demo.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/inference/cli_vae_demo.py b/inference/cli_vae_demo.py index 508aedb..cb95d82 100644 --- a/inference/cli_vae_demo.py +++ b/inference/cli_vae_demo.py @@ -143,7 +143,7 @@ if __name__ == "__main__": encoded_output = encode_video(args.model_path, args.video_path, dtype, device) torch.save(encoded_output, args.output_path + "/encoded.pt") print( - f"Finished encoding the video to a tensor, save it to a file at {encoded_output}/encoded.pt" + f"Finished encoding the video to a tensor, saved to {args.output_path}/encoded.pt" ) elif args.mode == "decode": assert args.encoded_path, "Encoded tensor path must be provided for decoding."