Connect with us

AI

Andersonの視点

AI

mm
ChatGPT/Firefly image depicting a jet-skier impossibly leaving a wake in front of himself.

2022 generative AI 2025 China generative Tencent Hunyuan Video hobbyist AI open-source full-world video diffusion model tailor to their needs .

Alibaba Wan 2.1 image-to-video FOSS customization Wan LoRAs .

human-centric foundation model SkyReels Alibaba VACE video creation editing suite :

Alibaba VACE Source: https://ali-vilab.github.io/VACE-Page/

generative video AI Arxiv Computer Vision 350 .

Stable Diffusion 2022 Dreambooth LoRA .

Video diffusion models Hunyuan Wan 2.1 temporal consistency humans environments objects .

VFX face-swapping ControlNet .

rock

Wan 2.1 prompt ‘ Source: https://videophy2.github.io/

text-to-video image-to-video physics bloopers .

theory Alibaba UAE models train single images videos training purposes ; temporal order .

data augmentation routines source training clip model forwards backwards .

2019 University of Bristol equivariant invariant irreversible source data video clips dataset .

Examples of three types of movement, only one of which is freely reversible while maintaining plausible physical dynamics. Source: https://arxiv.org/abs/1909.09422

Examples of three types of movement, only one of which is freely reversible while maintaining plausible physical dynamics. Source: https://arxiv.org/abs/1909.09422

Hunyuan Video Wan 2.1 arbitrarily ‘reversed’ clips training .

reports practical experience hyperscale datasets clips movements occurring in reverse .

rock

OpenAI Sora

VideoPhy benchmarking system GitHub .

Scores obtained by leading open and closed-source systems, with the output of the frameworks evaluated by human annotators. Source: https://arxiv.org/pdf/2503.06800

Scores obtained by leading open and closed-source systems, with the output of the frameworks evaluated by human annotators. Source: https://arxiv.org/pdf/2503.06800

VideoPhy-2 action-centric physical commonsense evaluation video generation .

UCLA Google Research VideoPhy-2: A Challenging Action-Centric Physical Commonsense Evaluation in Video Generation .

study project site code GitHub dataset viewer Hugging Face .

OpenAI Sora

VideoPhy-2 challenging commonsense evaluation dataset real-world actions .

LLM prompts actions physical activities object interactions .

Above: A text prompt is generated from an action using an LLM and used to create a video with a text-to-video generator. A vision-language model captions the video, identifying possible physical rules at play. Below: Human annotators evaluate the video’s realism, confirm rule violations, add missing rules, and check whether the video matches the original prompt.

Above: A text prompt is generated from an action using an LLM and used to create a video with a text-to-video generator. A vision-language model captions the video, identifying possible physical rules at play. Below: Human annotators evaluate the video’s realism, confirm rule violations, add missing rules, and check whether the video matches the original prompt.

Kinetics UCF-101 SSv2 datasets activities sports object interactions real-world physics .

STEM-trained student annotators minimum undergraduate qualification reviewed filtered list .

actions principles gravity momentum elasticity .

Gemini-2.0-Flash-Exp prompts actions dataset .

Samples from the distilled actions.

Samples from the distilled actions.

Sample prompts from VideoPhy-2, categorized by physical activities or object interactions. Each prompt is paired with its corresponding action and the relevant physical principle it tests.

Sample prompts from VideoPhy-2, categorized by physical activities or object interactions. Each prompt is paired with its corresponding action and the relevant physical principle it tests.

Examples from the upsampled captions.

Examples from the upsampled captions.

criteria videos input prompt physical principles .

rating-based feedback successes failures .

human annotators scored videos five-point scale .

The interface presented to the AMT annotators.

The interface presented to the AMT annotators.

CogVideoX-5B VideoCrafter2 HunyuanVideo-13B Cosmos-Diffusion Wan2.1-14B OpenAI Sora Luma Ray .

models prompted upsampled captions .

videos less than 6 seconds .

Results from the initial round.

Results from the initial round.

The top row shows videos generated by Wan2.1. (a) In Ray2, the jet-ski on the left lags behind before moving backward. (b) In Hunyuan-13B, the sledgehammer deforms mid-swing, and a broken wooden board appears unexpectedly. (c) In Cosmos-7B, the javelin expels sand before making contact with the ground.

The top row shows videos generated by Wan2.1. (a) In Ray2, the jet-ski on the left lags behind before moving backward. (b) In Hunyuan-13B, the sledgehammer deforms mid-swing, and a broken wooden board appears unexpectedly. (c) In Cosmos-7B, the javelin expels sand before making contact with the ground.

prompt ‘

prompt ‘

*

2025 3 13
The translation is provided while maintaining the exact same structure as the input, without adding or removing any information, and keeping all URLs, HTML tags, and special characters unchanged.

機械学習に関するライター、ヒューマンイメージシンセシスのドメインスペシャリスト。Metaphysic.aiの研究コンテンツ責任者を務めた。