Cora: Correspondence-aware image editing using few step diffusion

Cora is a novel few-step diffusion framework for image editing that leverages correspondence-aware noise correction and interpolated attention maps to effectively preserve structural attributes and identity while enabling significant non-rigid deformations, object modifications, and content generation with minimal artifacts.

Amirhossein Alimohammadi, Aryan Mikaeili, Sauradip Nag, Negar Hassanpour, Andrea Tagliasacchi, Ali Mahdavi-Amiri

Published 2026-03-02
📖 4 min read☕ Coffee break read

Imagine you have a photo of your friend sitting on a park bench. You want to use AI to edit the photo so that your friend is now jumping in the air, wearing a superhero cape, and the background has changed from a park to a city skyline.

Doing this with older AI tools is like trying to edit a clay sculpture by just painting over it. If you tell the AI to "make them jump," the AI might try to stretch the clay, resulting in a weird, melted face or a body that looks like it's made of rubber. If you ask it to add a cape, it might accidentally paint the cape onto the bench or make the friend's face disappear.

Cora is a new, smarter way to do this. Think of it as a masterful digital tailor who understands not just what the photo looks like, but how every single part of it connects to the others.

Here is how Cora works, broken down into three simple concepts:

1. The "Map" Problem (Correspondence-Aware Noise)

The Old Way: Imagine you have a map of a city. You decide to move the library to a new street. If you just drag the library icon on the map without updating the roads, the library ends up floating in the middle of a river.
The Cora Way: Cora creates a dynamic map (called "correspondence") before it starts editing. It looks at the original photo and the new idea and says, "Okay, the friend's left leg in the original photo is now in the air in the new photo. I need to move the texture of that leg to the new spot."
It ensures that when the AI "paints" the new image, it knows exactly where the original skin, hair, and clothes should go, even if they have moved or changed shape. This prevents the "melting" or "glitching" artifacts you see in other tools.

2. The "Blender" Problem (Attention Interpolation)

The Old Way: Imagine you are mixing two smoothies. One is strawberry (the original photo) and one is blueberry (the new idea).

  • Method A (Just copying): You only use the strawberry smoothie. The result tastes like strawberry, even though you wanted blueberry. The AI refuses to change the image enough.
  • Method B (Dumping them together): You dump the whole strawberry smoothie into the blueberry one. Now you have a chunky, weird mess where strawberry seeds are floating in blueberry juice. The AI accidentally puts parts of the original background onto the new object.
    The Cora Way: Cora uses a smart blender (called "Spherical Interpolation"). It doesn't just mix the two; it blends them perfectly based on how similar the ingredients are.
  • If the AI needs to keep the friend's face (because it's still the same person), it blends in the "strawberry" flavor.
  • If the AI needs to create a brand-new superhero cape that didn't exist before, it knows to use 100% "blueberry" (the new idea) and ignore the strawberry.
    This allows the AI to keep the identity of the person while inventing new things around them without them bleeding into each other.

3. The "Skeleton" Problem (Structural Alignment)

The Old Way: Imagine you are rearranging furniture in a room. If you just move the sofa without checking the walls, you might end up with the sofa floating in mid-air or blocking the door.
The Cora Way: Cora first checks the skeleton of the image. It asks, "Where are the main structures?" (e.g., the horizon line, the position of the person's head, the legs).
It locks these structures in place first. Then, it lets the "flesh" (the colors and textures) change. This ensures that even if your friend is jumping, they are still standing on the ground (or the air, if that's the prompt) in a way that makes physical sense, rather than floating randomly.

The Result?

Because Cora is built on a "fast-forward" version of AI (called Few-Step Diffusion), it does all this thinking and painting in just 4 steps instead of the usual 20–50 steps.

  • Old AI: Takes a long time, and the result looks like a melted wax figure.
  • Cora: Works instantly, keeps your friend looking like your friend, adds the cape perfectly, and makes the jump look natural.

In short: Cora is the difference between a clumsy painter who smears the canvas and a skilled editor who knows exactly which pixels belong to the "old story" and which pixels belong to the "new story," blending them together seamlessly.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →