Operator Reference

inpainting_ctinpainting_ctInpaintingCtInpaintingCtinpainting_ct (Operator)

inpainting_ctinpainting_ctInpaintingCtInpaintingCtinpainting_ct — Perform an inpainting by coherence transport.

Signature

Herror inpainting_ct(const Hobject Image, const Hobject Region, Hobject* InpaintedImage, double Epsilon, double Kappa, double Sigma, double Rho, double ChannelCoefficients)

Herror T_inpainting_ct(const Hobject Image, const Hobject Region, Hobject* InpaintedImage, const Htuple Epsilon, const Htuple Kappa, const Htuple Sigma, const Htuple Rho, const Htuple ChannelCoefficients)

void InpaintingCt(const HObject& Image, const HObject& Region, HObject* InpaintedImage, const HTuple& Epsilon, const HTuple& Kappa, const HTuple& Sigma, const HTuple& Rho, const HTuple& ChannelCoefficients)

HImage HImage::InpaintingCt(const HRegion& Region, double Epsilon, double Kappa, double Sigma, double Rho, const HTuple& ChannelCoefficients) const

HImage HImage::InpaintingCt(const HRegion& Region, double Epsilon, double Kappa, double Sigma, double Rho, double ChannelCoefficients) const

static void HOperatorSet.InpaintingCt(HObject image, HObject region, out HObject inpaintedImage, HTuple epsilon, HTuple kappa, HTuple sigma, HTuple rho, HTuple channelCoefficients)

HImage HImage.InpaintingCt(HRegion region, double epsilon, double kappa, double sigma, double rho, HTuple channelCoefficients)

HImage HImage.InpaintingCt(HRegion region, double epsilon, double kappa, double sigma, double rho, double channelCoefficients)

def inpainting_ct(image: HObject, region: HObject, epsilon: float, kappa: float, sigma: float, rho: float, channel_coefficients: MaybeSequence[float]) -> HObject

Description

The operator inpainting_ctinpainting_ctInpaintingCtInpaintingCtinpainting_ct inpaints a missing region RegionRegionRegionregionregion of an image ImageImageImageimageimage by transporting image information from the region's boundary along the coherence direction into this region.

Since this operator's basic concept is inpainting by continuing broken contour lines, the image content and inpainting region must be such that this idea makes sense. That is, if a contour line hits the region to inpaint at a pixel p, there should be some opposite point q where this contour line continues so that the continuation of contour lines from two opposite sides can succeed. In cases where there is less geometry in the image, a diffusion-based inpainter, e.g., harmonic_interpolationharmonic_interpolationHarmonicInterpolationHarmonicInterpolationharmonic_interpolation may yield better results. Alternatively, KappaKappaKappakappakappa can be set to 0. An extreme situation with little global geometries are pure textures. Then the idea behind this operator will fail to produce good results (think of a checkerboard with a big region to inpaint relative to the checker fields). For these kinds of images, a texture-based inpainting, e.g., inpainting_textureinpainting_textureInpaintingTextureInpaintingTextureinpainting_texture, can be used instead.

The operator uses a so-called upwind scheme to assign gray values to the missing pixels, i.e.,:

  • The order of the pixels to process is given by their Euclidean distance to the boundary of the region to inpaint.

  • A new value is computed as a weighted average of already known values within a disc of radius EpsilonEpsilonEpsilonepsilonepsilon around the current pixel. The disc is restricted to already known pixels.

  • The size of this scheme's mask depends on EpsilonEpsilonEpsilonepsilonepsilon.

The initially used image data comes from a stripe of thickness EpsilonEpsilonEpsilonepsilonepsilon around the region to inpaint. Thus, EpsilonEpsilonEpsilonepsilonepsilon must be at least 1 for the scheme to work, but should be greater. The maximum value for EpsilonEpsilonEpsilonepsilonepsilon depends on the gray values that should be transported into the region. Choosing EpsilonEpsilonEpsilonepsilonepsilon = 5 can be used in many cases.

Since the goal is to close broken contour lines, the direction of the level lines must be estimated and used in the weight. This estimated direction is called the coherence direction, and is computed by means of the structure tensor S. and where * denotes the convolution, u denotes the gray value image, D the derivative and G Gaussian kernels with standard deviation and . These standard deviations are defined by the operator's parameters SigmaSigmaSigmasigmasigma and RhoRhoRhorhorho. SigmaSigmaSigmasigmasigma should have the size of the noise or unimportant little objects, which are then not considered in the estimation step by the pre-smoothing. RhoRhoRhorhorho gives the size of the window around a pixel that will be used for direction estimation. The coherence direction c then is given by the eigendirection of S with respect to the minimal eigenvalue , i.e.

For multichannel or color images, the scheme above is applied to each channel separately, but the weights must be the same for all channels to propagate information in the same direction. Since the weight depends on the coherence direction, the common direction is given by the eigendirection of a composite structure tensor. If denote the n channels of the image, the channel structure tensors are computed and then combined to the composite structure tensor S. The coefficients are passed in ChannelCoefficientsChannelCoefficientsChannelCoefficientschannelCoefficientschannel_coefficients, which is a tuple of length n or length 1. If the tuple length is 1, the arithmetic mean is used, i.e., . If the length of ChannelCoefficientsChannelCoefficientsChannelCoefficientschannelCoefficientschannel_coefficients matches the number of channels, the are set to in order to get a well-defined convex combination. Hence, the ChannelCoefficientsChannelCoefficientsChannelCoefficientschannelCoefficientschannel_coefficients must be greater than or equal to zero and their sum must be greater than zero. If the tuple length is neither 1 nor the number of channels or the requirement above is not satisfied, the operator returns an error message.

The purpose of using other ChannelCoefficientsChannelCoefficientsChannelCoefficientschannelCoefficientschannel_coefficients than the arithmetic mean is to adapt to different color codes. The coherence direction is a geometrical information of the composite image, which is given by high contrasts such as edges. Thus the more contrast a channel has, the more geometrical information it contains, and consequently the greater its coefficient should be chosen (relative to the others). For RGB images, [0.299, 0.587, 0.114] is a good choice.

The weight in the scheme is the product of a directional component and a distance component. If p is the 2D coordinate vector of the current pixel to be inpainted and q the 2D coordinate of a pixel in the neighborhood (the disc restricted to already known pixels), the directional component measures the deviation of the vector p-q from the coherence direction. If the deviation exponentially scaled by is large, a low directional component is assigned, whereas if it is small, a large directional component is assigned. is controlled by KappaKappaKappakappakappa (in percent): KappaKappaKappakappakappa defines how important it is to propagate information along the coherence direction, so a large KappaKappaKappakappakappa yields sharp edges, while a low KappaKappaKappakappakappa allows for more diffusion.

A special case is when KappaKappaKappakappakappa is zero: In this case the directional component of the weight is constant (one). The direction estimation step is then skipped to save computational costs and the parameters SigmaSigmaSigmasigmasigma, RhoRhoRhorhorho, ChannelCoefficientsChannelCoefficientsChannelCoefficientschannelCoefficientschannel_coefficients become meaningless, i.e, the propagation of information is not based on the structures visible in the image.

The distance component is 1/|p-q|. Consequently, if q is far away from p, a low distance component is assigned, whereas if it is near to p, a high distance component is assigned.

Attention

Note that filter operators may return unexpected results if an image with a reduced domain is used as input. Please refer to the chapter Filters.

Execution Information

  • Multithreading type: reentrant (runs in parallel with non-exclusive operators).
  • Multithreading scope: global (may be called from any thread).
  • Automatically parallelized on tuple level.

Parameters

ImageImageImageimageimage (input_object)  (multichannel-)image(-array) objectHImageHObjectHObjectHobject (byte / uint2 / real)

Input image.

RegionRegionRegionregionregion (input_object)  region objectHRegionHObjectHObjectHobject

Inpainting region.

InpaintedImageInpaintedImageInpaintedImageinpaintedImageinpainted_image (output_object)  (multichannel-)image(-array) objectHImageHObjectHObjectHobject * (byte / uint2 / real)

Output image.

EpsilonEpsilonEpsilonepsilonepsilon (input_control)  number HTuplefloatHTupleHtuple (real) (double) (double) (double)

Radius of the pixel neighborhood.

Default: 5.0

Value range: 1.0 ≤ Epsilon Epsilon Epsilon epsilon epsilon ≤ 20.0

Minimum increment: 1.0

Recommended increment: 1.0

KappaKappaKappakappakappa (input_control)  number HTuplefloatHTupleHtuple (real) (double) (double) (double)

Sharpness parameter in percent.

Default: 25.0

Value range: 0.0 ≤ Kappa Kappa Kappa kappa kappa ≤ 100.0

Minimum increment: 1.0

Recommended increment: 1.0

SigmaSigmaSigmasigmasigma (input_control)  number HTuplefloatHTupleHtuple (real) (double) (double) (double)

Pre-smoothing parameter.

Default: 1.41

Value range: 0.0 ≤ Sigma Sigma Sigma sigma sigma ≤ 20.0

Minimum increment: 0.001

Recommended increment: 0.01

RhoRhoRhorhorho (input_control)  number HTuplefloatHTupleHtuple (real) (double) (double) (double)

Smoothing parameter for the direction estimation.

Default: 4.0

Value range: 0.001 ≤ Rho Rho Rho rho rho ≤ 20.0

Minimum increment: 0.001

Recommended increment: 0.01

ChannelCoefficientsChannelCoefficientsChannelCoefficientschannelCoefficientschannel_coefficients (input_control)  number(-array) HTupleMaybeSequence[float]HTupleHtuple (real) (double) (double) (double)

Channel weights.

Default: 1

Example (HDevelop)

read_image (Image, 'claudia')
gen_circle (Circle, 333, 164, 35)
inpainting_ct (Image, Circle, InpaintedImage, 15, 25, 1.5, 3,1.0)

Alternatives

harmonic_interpolationharmonic_interpolationHarmonicInterpolationHarmonicInterpolationharmonic_interpolation, inpainting_anisoinpainting_anisoInpaintingAnisoInpaintingAnisoinpainting_aniso, inpainting_mcfinpainting_mcfInpaintingMcfInpaintingMcfinpainting_mcf, inpainting_cedinpainting_cedInpaintingCedInpaintingCedinpainting_ced, inpainting_textureinpainting_textureInpaintingTextureInpaintingTextureinpainting_texture

References

Folkmar Bornemann, Tom März: “Fast Image Inpainting Based On Coherence Transport”; Journal of Mathematical Imaging and Vision; vol. 28, no. 3; pp. 259-278; 2007.

Module

Foundation