Midjourney’s newly added “Model Personalization” turns “your aesthetic preferences” into a callable setting, making the same set of prompts more stable across different projects and more like you. This article focuses only on the core logic of this new feature, how to enable it, and how to incorporate it into a reusable image-generation workflow.
What is Midjourney’s personalization model: teaching the model your preferences
The essence of personalization is that Midjourney generates a “style inclination” based on your feedback about which images you prefer. After that, when you generate images in Midjourney, the model will be more inclined to output the compositions, textures, and aesthetic directions you tend to choose.
It’s better at unifying style than at “remembering specific people or assets” for you. If you want to create brand visuals, a poster series, or a consistent illustration style, Midjourney personalization can noticeably save the time you’d otherwise spend repeatedly revising prompts.
How to enable it: two entry points—Web and Discord
After logging in on the Midjourney website, you can find the personalization toggle on the relevant settings page, which will guide you through preference training (usually by choosing/ranking images). After providing a certain amount of preference feedback, the personalization effect will gradually stabilize.
If you’re used to working in Discord, you can also enable the personalization option through the usual settings-command entry (for example, opening the settings panel). No matter where you enable it, the core is to let Midjourney obtain your preference data and apply it to subsequent generations.


