In this paper, we extend the study of concept ablation within pre-trained models as introduced in ‘Ablating Concepts in Text-to-Image Diffusion Models’. Our work focuses on reproducing the results achieved by the different variants of concept ablation proposed through predefined metrics. We also introduce a novel variant of concept ablation—trademark ablation. This variant combines the principles of memorization and instance ablation to tackle the nuanced influence of proprietary or branded elements in model outputs. Further, our research contributions include an observational analysis of the model’s limitations. Moreover, we investigate the model’s behavior in response to ablation leakage-inducing prompts, which aim to indirectly ablate concepts, revealing insights into the model’s resilience and adaptability. We also observe the model’s performance degradation on images generated by concepts far from its target ablation concept.