Anderson's Angle
Now NSFW and ‘Celebrity’ Poses Are Fodder for AI Censorship

A new AI safeguard for generative video systems proposes censoring body poses. Physical stances (or facial expressions) that may be interpreted as sexually suggestive, ‘offensive gestures', or even copyrighted celebrity or potentially trademarked poses, are all targeted.
New research from China and Singapore addresses one of the less obvious domains in ‘unsafe' image and video generation: the depiction of a pose itself, in the sense of the disposition of the body or facial expression of a depicted person in AI-created output:

Conceptual schema for PoseGuard, the system proposed in the new research. Source: https://arxiv.org/pdf/2508.02476
The system, titled PoseGuard, uses fine-tuning and LoRAs to create models that intrinsically cannot generate ‘banned' poses. This approach was taken because the safeguards built into FOSS models can usually be trivially overcome, emphasizing that this new ‘filter' specifically targets local installations (since API-only models can filter inbound and outbound content and prompts, without the need to imperil the integrity of the model weights by fine-tuning).
This is not the first work to treat poses as unsafe data in themselves; ‘sexual facial expressions' have been a minor sub-field of study for some time, while several of the authors of the new work also created the less sophisticated Dormant system.
However, the new paper is the first, as far as I can tell, to extend the typing of poses beyond sexual content, even to the point of including ‘copyrighted celebrity movements':
‘We define unsafe poses based on the potential risks of generated outputs rather than geometric characteristics. [Unsafe] poses include: 1) discriminatory poses (e.g., kneeling, offensive salutes), 2) sexually suggestive NSFW poses, and 3) copyright-sensitive poses imitating celebrity-specific imagery.
‘These poses are collected through online sources (e.g., Wikipedia), LLM-based filtering, and risk-labeled datasets (e.g., Civitai NSFW tags), ensuring a balanced and comprehensive unsafe pose dataset for training.'

The ‘NSFW' category of the core 50 poses developed for PoseGuard.
It is interesting to note that celebrity poses can be trademarked or protected by legal means, and that adequately ‘creative' combinations of poses or stances can be protected as unique sequences of choreography. However, even an iconic single pose may not be protected, as one photographer discovered, in the Rentmeester Vs. Nike ruling:

A photographer who took the leftmost photo of Michael Jordan sued Nike when they recreated the photo (right); however, a panel of judges rejected the claim. Source: https://writtendescription.blogspot.com/2018/02/can-you-copyright-pose.html
The new PoseGuard system claims to be the first to degrade output when an unsafe pose is detected; to embed safety guardrails directly into a generative model; to define ‘unsafe' poses across three categories; and to ensure that generation retains quality and integrity once an offending pose has been altered enough to escape the filter.
The new paper is titled PoseGuard: Pose-Guided Generation with Safety Guardrails, and comes from six researchers across the University of Science and Technology of China, the (Singaporean) Agency for Science, Technology and Research (A*STAR CFAR), and Nanyang Technological University.
Method
PoseGuard repurposes the logic of backdoor attacks to build a defense mechanism directly into the model. In a typical backdoor attack, specific inputs trigger malicious outputs, and PoseGuard inverts this setup: certain predefined poses that are deemed unsafe due to their sexual, offensive, or copyright-sensitive nature, are linked to ‘neutral' target images, such as blank or blurred frames.
By fine-tuning the model on a combined dataset of normal and trigger poses, the system learns to preserve fidelity for benign inputs while degrading output quality for unsafe ones:

PoseGuard processes a reference image and pose sequence using a shared denoising UNet, combining pretrained weights with safety-aligned fine-tuning. This setup allows the model to suppress harmful generations from unsafe poses while maintaining output quality for normal inputs.
This ‘in-model' strategy eliminates the need for external filters, and remains effective even in adversarial or open-source environments.*
Data and Tests
To obtain benign baseline poses, the authors used the UBC-Fashion dataset:

Examples from the University of British Columbia fashion dataset, used as a source of benign poses in PoseGuard. Abstract poses were extracted from these images with a pose-estimation framework. Source: https://www.cs.ubc.ca/~lsigal/Publications/bmvc2019zablotskaia.pdf
Unsafe poses, as mentioned earlier, were sourced from open-source platforms such as CivitAI. Poses were extracted using the DWPose framework, resulting in 768x768px pose images:

Examples from the 50 unsafe poses used in training. Shown here are NSFW and copyright-sensitive poses, sourced from Wikipedia, Render-State, Civitai, and Google Search.
The pose-guided generation model was AnimateAnyone.
The six metrics used were FrĂ©chet Video Distance (FVD); FID-VID; Structural Similarity Index (SSIM); Peak Signal-to-Noise Ratio (PSNR); Learned Perceptual Similarity Metrics (LPIPS); and FrĂ©chet Inception Distance (FID). Tests were conducted on a NVIDIA A6000 GPU with 48GB of VRAM, at a batch size of 4 and a learning rate of 1×10-5.
The three primary categories tested for were effectiveness, robustness, and generalization.
In the first of these, effectiveness, the authors compared two training strategies for PoseGuard: full fine-tuning of the denoising UNet and parameter-efficient fine-tuning using LoRA modules.
Both approaches suppress outputs from unsafe poses while preserving output quality on benign poses, but with different trade-offs: full fine-tuning achieves stronger suppression and maintains higher fidelity, particularly when the number of unsafe training poses was small; and LoRA-based tuning introduces more degradation in generation quality as the number of unsafe poses increases â but requires significantly fewer parameters, and less compute.

PoseGuard performance across generation and defense metrics. Upward arrows indicate metrics where higher values are better; downward arrows indicate metrics where lower values are better.
Qualitative results (see image below) showed that, without intervention, the model reproduced offensive and NSFW poses with high fidelity. With PoseGuard enabled, these poses triggered low-quality or blank outputs, while benign inputs remained visually intact. As the defense set grew from four to thirty-two unsafe poses, benign output quality declined moderately, especially for LoRA.

Visual results showing how PoseGuard responds to a single unsafe pose using full-parameter fine-tuning. The model suppresses output for discriminatory, NSFW, and copyright-sensitive poses, redirecting them to a black image, while preserving quality for normal inputs.
For robustness, PoseGuard was tested under conditions that simulate real-world deployment, where input poses may not match predefined examples exactly. The evaluation included common transformations such as translation, scaling, and rotation, as well as manual adjustments to joint angles to mimic natural variation.

Results for robustness of PoseGuard in the face of common pose transformations.
In most cases, the model continued to suppress unsafe generations, indicating that the defense remains robust to moderate perturbations. When the alterations removed the underlying risk in the pose, the model stopped suppressing and produced normal outputs, suggesting that it avoids false positives under benign deviations.

Evaluation of PoseGuardâs robustness to pose modifications. The figure shows model outputs for unsafe poses altered by translation, scaling, and rotation, as well as manual limb adjustments. PoseGuard continues to suppress unsafe generations under mild changes, but resumes normal output when the pose no longer carries ‘risky' content.
Finally, in the main run of experiments, the researchers tested PoseGuard for generalization â its ability to operate effectively on novel data, in a range of environments and circumstances.
Here, PoseGuard was applied to reference image-guided generation using the aforementioned AnimateAnyone model. In this setting, the system showed stronger suppression of unauthorized outputs compared to pose-based control, with near-total degradation of the generated video in some cases:

Comparison of PoseGuard's performance when applied to pose-guided versus reference image-guided generation, using full fine-tuning on four unsafe inputs.
The authors attribute this to the dense identity information in reference images, which allows the model to more easily learn targeted defensive behavior. The results, they suggest, indicate that PoseGuard can limit impersonation risks in scenarios where video is generated directly from a person's appearance.
For a final test, the authors applied PoseGuard to facial landmark-guided video synthesis using the AniPortrait system, a scenario that targets fine-grained facial expressions rather than full-body poses.

Unsafe facial expressions suppressed in AniPortrait, with the new system.
By fine-tuning the Denoising UNet with the same defense mechanism, the model was able to suppress outputs from unsafe facial landmarks while leaving benign expressions unaffected. The results, the authors suggest, show that PoseGuard can generalize across input modalities and maintain effectiveness in more localized, expression-driven generation tasks.

Visual results showing the way that PoseGuard responds to reference image-guided generation.
Conclusion
It has to be admitted that for many of the 50 banned reference poses supplied by the paper, activities such as medical examinations, or even doing dull housework tasks, would likely get blocked in what can only be conceived of as a synthesis-based version of the Scunthorpe effect.
From that standpoint, and much more so in the case of facial expressions, (which can be much more ambiguous and nuanced in intent), PoseGuard would seem to be something of a blunt instrument. To boot, due to a general chilling effect around NSFW AI, FOSS releases such as the recent Flux Kontext are routinely very censored in any case,, either through rigorous dataset filtering, weight-editing, or both.
Therefore adding the constrictions proposed here to the burden of local-model censorship seems like a tacit bid to suppress the effectiveness of non-API generative systems. This perhaps points us towards a future where local models can produce an inferior generation of anything the user likes, while API models offer infinitely superior output, if one can only negotiate the gauntlet of filters and safeguards that pacify the host company's legal department.
A system such as PoseGuard, wherein fine-tuning actively affects the quality of the base model's output (though this is overlooked in the paper), is not aimed at API systems at all; online only vanguard models will likely continue to benefit from unconstrained training data, since the formidable NSFW capacities of these models are reined in by considerable oversight measures.
* Method is as short here as in the paper (which runs to only five pages), and, as usual, the approach is best-understood from the tests section.
First published Wednesday, August 6, 2025