opencv-contrib-python-headless 4.13.0.90__cp37-abi3-macosx_14_0_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- cv2/.dylibs/libaom.3.13.1.dylib +0 -0
- cv2/.dylibs/libavif.16.3.0.dylib +0 -0
- cv2/.dylibs/libdav1d.7.dylib +0 -0
- cv2/.dylibs/libvmaf.3.dylib +0 -0
- cv2/Error/__init__.pyi +118 -0
- cv2/LICENSE-3RD-PARTY.txt +3513 -0
- cv2/LICENSE.txt +21 -0
- cv2/__init__.py +181 -0
- cv2/__init__.pyi +6858 -0
- cv2/aruco/__init__.pyi +410 -0
- cv2/barcode/__init__.pyi +40 -0
- cv2/bgsegm/__init__.pyi +202 -0
- cv2/bioinspired/__init__.pyi +121 -0
- cv2/ccm/__init__.pyi +167 -0
- cv2/colored_kinfu/__init__.pyi +96 -0
- cv2/config-3.py +24 -0
- cv2/config.py +5 -0
- cv2/cuda/__init__.pyi +553 -0
- cv2/cv2.abi3.so +0 -0
- cv2/data/__init__.py +3 -0
- cv2/data/haarcascade_eye.xml +12213 -0
- cv2/data/haarcascade_eye_tree_eyeglasses.xml +22619 -0
- cv2/data/haarcascade_frontalcatface.xml +14382 -0
- cv2/data/haarcascade_frontalcatface_extended.xml +13394 -0
- cv2/data/haarcascade_frontalface_alt.xml +24350 -0
- cv2/data/haarcascade_frontalface_alt2.xml +20719 -0
- cv2/data/haarcascade_frontalface_alt_tree.xml +96484 -0
- cv2/data/haarcascade_frontalface_default.xml +33314 -0
- cv2/data/haarcascade_fullbody.xml +17030 -0
- cv2/data/haarcascade_lefteye_2splits.xml +7390 -0
- cv2/data/haarcascade_license_plate_rus_16stages.xml +1404 -0
- cv2/data/haarcascade_lowerbody.xml +14056 -0
- cv2/data/haarcascade_profileface.xml +29690 -0
- cv2/data/haarcascade_righteye_2splits.xml +7407 -0
- cv2/data/haarcascade_russian_plate_number.xml +2656 -0
- cv2/data/haarcascade_smile.xml +6729 -0
- cv2/data/haarcascade_upperbody.xml +28134 -0
- cv2/datasets/__init__.pyi +80 -0
- cv2/detail/__init__.pyi +627 -0
- cv2/dnn/__init__.pyi +549 -0
- cv2/dnn_superres/__init__.pyi +37 -0
- cv2/dpm/__init__.pyi +10 -0
- cv2/dynafu/__init__.pyi +43 -0
- cv2/face/__init__.pyi +220 -0
- cv2/fisheye/__init__.pyi +88 -0
- cv2/flann/__init__.pyi +65 -0
- cv2/ft/__init__.pyi +98 -0
- cv2/gapi/__init__.py +323 -0
- cv2/gapi/__init__.pyi +349 -0
- cv2/gapi/core/__init__.pyi +7 -0
- cv2/gapi/core/cpu/__init__.pyi +9 -0
- cv2/gapi/core/fluid/__init__.pyi +9 -0
- cv2/gapi/core/ocl/__init__.pyi +9 -0
- cv2/gapi/ie/__init__.pyi +51 -0
- cv2/gapi/ie/detail/__init__.pyi +12 -0
- cv2/gapi/imgproc/__init__.pyi +5 -0
- cv2/gapi/imgproc/fluid/__init__.pyi +9 -0
- cv2/gapi/oak/__init__.pyi +37 -0
- cv2/gapi/onnx/__init__.pyi +55 -0
- cv2/gapi/onnx/ep/__init__.pyi +63 -0
- cv2/gapi/ot/__init__.pyi +32 -0
- cv2/gapi/ot/cpu/__init__.pyi +9 -0
- cv2/gapi/ov/__init__.pyi +74 -0
- cv2/gapi/own/__init__.pyi +5 -0
- cv2/gapi/own/detail/__init__.pyi +10 -0
- cv2/gapi/render/__init__.pyi +5 -0
- cv2/gapi/render/ocv/__init__.pyi +9 -0
- cv2/gapi/streaming/__init__.pyi +42 -0
- cv2/gapi/video/__init__.pyi +10 -0
- cv2/gapi/wip/__init__.pyi +43 -0
- cv2/gapi/wip/draw/__init__.pyi +119 -0
- cv2/gapi/wip/gst/__init__.pyi +17 -0
- cv2/gapi/wip/onevpl/__init__.pyi +16 -0
- cv2/hfs/__init__.pyi +53 -0
- cv2/img_hash/__init__.pyi +116 -0
- cv2/instr/__init__.pyi +24 -0
- cv2/intensity_transform/__init__.pyi +27 -0
- cv2/ipp/__init__.pyi +14 -0
- cv2/kinfu/__init__.pyi +133 -0
- cv2/kinfu/detail/__init__.pyi +7 -0
- cv2/large_kinfu/__init__.pyi +73 -0
- cv2/legacy/__init__.pyi +93 -0
- cv2/line_descriptor/__init__.pyi +112 -0
- cv2/linemod/__init__.pyi +151 -0
- cv2/load_config_py2.py +6 -0
- cv2/load_config_py3.py +9 -0
- cv2/mat_wrapper/__init__.py +40 -0
- cv2/mcc/__init__.pyi +109 -0
- cv2/misc/__init__.py +1 -0
- cv2/misc/version.py +5 -0
- cv2/ml/__init__.pyi +696 -0
- cv2/motempl/__init__.pyi +29 -0
- cv2/multicalib/__init__.pyi +10 -0
- cv2/ocl/__init__.pyi +252 -0
- cv2/ogl/__init__.pyi +51 -0
- cv2/omnidir/__init__.pyi +68 -0
- cv2/optflow/__init__.pyi +286 -0
- cv2/parallel/__init__.pyi +6 -0
- cv2/phase_unwrapping/__init__.pyi +41 -0
- cv2/plot/__init__.pyi +64 -0
- cv2/ppf_match_3d/__init__.pyi +91 -0
- cv2/py.typed +0 -0
- cv2/quality/__init__.pyi +149 -0
- cv2/rapid/__init__.pyi +91 -0
- cv2/reg/__init__.pyi +210 -0
- cv2/rgbd/__init__.pyi +449 -0
- cv2/saliency/__init__.pyi +117 -0
- cv2/samples/__init__.pyi +12 -0
- cv2/segmentation/__init__.pyi +39 -0
- cv2/signal/__init__.pyi +14 -0
- cv2/stereo/__init__.pyi +88 -0
- cv2/structured_light/__init__.pyi +94 -0
- cv2/text/__init__.pyi +204 -0
- cv2/typing/__init__.py +180 -0
- cv2/utils/__init__.py +14 -0
- cv2/utils/__init__.pyi +110 -0
- cv2/utils/fs/__init__.pyi +6 -0
- cv2/utils/logging/__init__.pyi +22 -0
- cv2/utils/nested/__init__.pyi +31 -0
- cv2/version.py +5 -0
- cv2/videoio_registry/__init__.pyi +31 -0
- cv2/videostab/__init__.pyi +16 -0
- cv2/wechat_qrcode/__init__.pyi +23 -0
- cv2/xfeatures2d/__init__.pyi +537 -0
- cv2/ximgproc/__init__.pyi +746 -0
- cv2/ximgproc/segmentation/__init__.pyi +116 -0
- cv2/xphoto/__init__.pyi +142 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/LICENSE-3RD-PARTY.txt +3513 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/LICENSE.txt +21 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/METADATA +300 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/RECORD +133 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/WHEEL +6 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/top_level.txt +1 -0
|
@@ -0,0 +1,116 @@
|
|
|
1
|
+
__all__: list[str] = []
|
|
2
|
+
|
|
3
|
+
import cv2
|
|
4
|
+
import cv2.typing
|
|
5
|
+
import typing as _typing
|
|
6
|
+
|
|
7
|
+
|
|
8
|
+
# Classes
|
|
9
|
+
class GraphSegmentation(cv2.Algorithm):
|
|
10
|
+
# Functions
|
|
11
|
+
@_typing.overload
|
|
12
|
+
def processImage(self, src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
13
|
+
@_typing.overload
|
|
14
|
+
def processImage(self, src: cv2.UMat, dst: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
15
|
+
|
|
16
|
+
def setSigma(self, sigma: float) -> None: ...
|
|
17
|
+
|
|
18
|
+
def getSigma(self) -> float: ...
|
|
19
|
+
|
|
20
|
+
def setK(self, k: float) -> None: ...
|
|
21
|
+
|
|
22
|
+
def getK(self) -> float: ...
|
|
23
|
+
|
|
24
|
+
def setMinSize(self, min_size: int) -> None: ...
|
|
25
|
+
|
|
26
|
+
def getMinSize(self) -> int: ...
|
|
27
|
+
|
|
28
|
+
|
|
29
|
+
class SelectiveSearchSegmentationStrategy(cv2.Algorithm):
|
|
30
|
+
# Functions
|
|
31
|
+
@_typing.overload
|
|
32
|
+
def setImage(self, img: cv2.typing.MatLike, regions: cv2.typing.MatLike, sizes: cv2.typing.MatLike, image_id: int = ...) -> None: ...
|
|
33
|
+
@_typing.overload
|
|
34
|
+
def setImage(self, img: cv2.UMat, regions: cv2.UMat, sizes: cv2.UMat, image_id: int = ...) -> None: ...
|
|
35
|
+
|
|
36
|
+
def get(self, r1: int, r2: int) -> float: ...
|
|
37
|
+
|
|
38
|
+
def merge(self, r1: int, r2: int) -> None: ...
|
|
39
|
+
|
|
40
|
+
|
|
41
|
+
class SelectiveSearchSegmentationStrategyColor(SelectiveSearchSegmentationStrategy):
|
|
42
|
+
...
|
|
43
|
+
|
|
44
|
+
class SelectiveSearchSegmentationStrategySize(SelectiveSearchSegmentationStrategy):
|
|
45
|
+
...
|
|
46
|
+
|
|
47
|
+
class SelectiveSearchSegmentationStrategyTexture(SelectiveSearchSegmentationStrategy):
|
|
48
|
+
...
|
|
49
|
+
|
|
50
|
+
class SelectiveSearchSegmentationStrategyFill(SelectiveSearchSegmentationStrategy):
|
|
51
|
+
...
|
|
52
|
+
|
|
53
|
+
class SelectiveSearchSegmentationStrategyMultiple(SelectiveSearchSegmentationStrategy):
|
|
54
|
+
# Functions
|
|
55
|
+
def addStrategy(self, g: SelectiveSearchSegmentationStrategy, weight: float) -> None: ...
|
|
56
|
+
|
|
57
|
+
def clearStrategies(self) -> None: ...
|
|
58
|
+
|
|
59
|
+
|
|
60
|
+
class SelectiveSearchSegmentation(cv2.Algorithm):
|
|
61
|
+
# Functions
|
|
62
|
+
@_typing.overload
|
|
63
|
+
def setBaseImage(self, img: cv2.typing.MatLike) -> None: ...
|
|
64
|
+
@_typing.overload
|
|
65
|
+
def setBaseImage(self, img: cv2.UMat) -> None: ...
|
|
66
|
+
|
|
67
|
+
def switchToSingleStrategy(self, k: int = ..., sigma: float = ...) -> None: ...
|
|
68
|
+
|
|
69
|
+
def switchToSelectiveSearchFast(self, base_k: int = ..., inc_k: int = ..., sigma: float = ...) -> None: ...
|
|
70
|
+
|
|
71
|
+
def switchToSelectiveSearchQuality(self, base_k: int = ..., inc_k: int = ..., sigma: float = ...) -> None: ...
|
|
72
|
+
|
|
73
|
+
@_typing.overload
|
|
74
|
+
def addImage(self, img: cv2.typing.MatLike) -> None: ...
|
|
75
|
+
@_typing.overload
|
|
76
|
+
def addImage(self, img: cv2.UMat) -> None: ...
|
|
77
|
+
|
|
78
|
+
def clearImages(self) -> None: ...
|
|
79
|
+
|
|
80
|
+
def addGraphSegmentation(self, g: GraphSegmentation) -> None: ...
|
|
81
|
+
|
|
82
|
+
def clearGraphSegmentations(self) -> None: ...
|
|
83
|
+
|
|
84
|
+
def addStrategy(self, s: SelectiveSearchSegmentationStrategy) -> None: ...
|
|
85
|
+
|
|
86
|
+
def clearStrategies(self) -> None: ...
|
|
87
|
+
|
|
88
|
+
def process(self) -> _typing.Sequence[cv2.typing.Rect]: ...
|
|
89
|
+
|
|
90
|
+
|
|
91
|
+
|
|
92
|
+
# Functions
|
|
93
|
+
def createGraphSegmentation(sigma: float = ..., k: float = ..., min_size: int = ...) -> GraphSegmentation: ...
|
|
94
|
+
|
|
95
|
+
def createSelectiveSearchSegmentation() -> SelectiveSearchSegmentation: ...
|
|
96
|
+
|
|
97
|
+
def createSelectiveSearchSegmentationStrategyColor() -> SelectiveSearchSegmentationStrategyColor: ...
|
|
98
|
+
|
|
99
|
+
def createSelectiveSearchSegmentationStrategyFill() -> SelectiveSearchSegmentationStrategyFill: ...
|
|
100
|
+
|
|
101
|
+
@_typing.overload
|
|
102
|
+
def createSelectiveSearchSegmentationStrategyMultiple() -> SelectiveSearchSegmentationStrategyMultiple: ...
|
|
103
|
+
@_typing.overload
|
|
104
|
+
def createSelectiveSearchSegmentationStrategyMultiple(s1: SelectiveSearchSegmentationStrategy) -> SelectiveSearchSegmentationStrategyMultiple: ...
|
|
105
|
+
@_typing.overload
|
|
106
|
+
def createSelectiveSearchSegmentationStrategyMultiple(s1: SelectiveSearchSegmentationStrategy, s2: SelectiveSearchSegmentationStrategy) -> SelectiveSearchSegmentationStrategyMultiple: ...
|
|
107
|
+
@_typing.overload
|
|
108
|
+
def createSelectiveSearchSegmentationStrategyMultiple(s1: SelectiveSearchSegmentationStrategy, s2: SelectiveSearchSegmentationStrategy, s3: SelectiveSearchSegmentationStrategy) -> SelectiveSearchSegmentationStrategyMultiple: ...
|
|
109
|
+
@_typing.overload
|
|
110
|
+
def createSelectiveSearchSegmentationStrategyMultiple(s1: SelectiveSearchSegmentationStrategy, s2: SelectiveSearchSegmentationStrategy, s3: SelectiveSearchSegmentationStrategy, s4: SelectiveSearchSegmentationStrategy) -> SelectiveSearchSegmentationStrategyMultiple: ...
|
|
111
|
+
|
|
112
|
+
def createSelectiveSearchSegmentationStrategySize() -> SelectiveSearchSegmentationStrategySize: ...
|
|
113
|
+
|
|
114
|
+
def createSelectiveSearchSegmentationStrategyTexture() -> SelectiveSearchSegmentationStrategyTexture: ...
|
|
115
|
+
|
|
116
|
+
|
cv2/xphoto/__init__.pyi
ADDED
|
@@ -0,0 +1,142 @@
|
|
|
1
|
+
__all__: list[str] = []
|
|
2
|
+
|
|
3
|
+
import cv2
|
|
4
|
+
import cv2.typing
|
|
5
|
+
import typing as _typing
|
|
6
|
+
|
|
7
|
+
|
|
8
|
+
# Enumerations
|
|
9
|
+
HAAR: int
|
|
10
|
+
TransformTypes = int
|
|
11
|
+
"""One of [HAAR]"""
|
|
12
|
+
|
|
13
|
+
BM3D_STEPALL: int
|
|
14
|
+
BM3D_STEP1: int
|
|
15
|
+
BM3D_STEP2: int
|
|
16
|
+
Bm3dSteps = int
|
|
17
|
+
"""One of [BM3D_STEPALL, BM3D_STEP1, BM3D_STEP2]"""
|
|
18
|
+
|
|
19
|
+
INPAINT_SHIFTMAP: int
|
|
20
|
+
INPAINT_FSR_BEST: int
|
|
21
|
+
INPAINT_FSR_FAST: int
|
|
22
|
+
InpaintTypes = int
|
|
23
|
+
"""One of [INPAINT_SHIFTMAP, INPAINT_FSR_BEST, INPAINT_FSR_FAST]"""
|
|
24
|
+
|
|
25
|
+
|
|
26
|
+
|
|
27
|
+
# Classes
|
|
28
|
+
class TonemapDurand(cv2.Tonemap):
|
|
29
|
+
# Functions
|
|
30
|
+
def getSaturation(self) -> float: ...
|
|
31
|
+
|
|
32
|
+
def setSaturation(self, saturation: float) -> None: ...
|
|
33
|
+
|
|
34
|
+
def getContrast(self) -> float: ...
|
|
35
|
+
|
|
36
|
+
def setContrast(self, contrast: float) -> None: ...
|
|
37
|
+
|
|
38
|
+
def getSigmaSpace(self) -> float: ...
|
|
39
|
+
|
|
40
|
+
def setSigmaSpace(self, sigma_space: float) -> None: ...
|
|
41
|
+
|
|
42
|
+
def getSigmaColor(self) -> float: ...
|
|
43
|
+
|
|
44
|
+
def setSigmaColor(self, sigma_color: float) -> None: ...
|
|
45
|
+
|
|
46
|
+
|
|
47
|
+
class WhiteBalancer(cv2.Algorithm):
|
|
48
|
+
# Functions
|
|
49
|
+
@_typing.overload
|
|
50
|
+
def balanceWhite(self, src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
51
|
+
@_typing.overload
|
|
52
|
+
def balanceWhite(self, src: cv2.UMat, dst: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
53
|
+
|
|
54
|
+
|
|
55
|
+
class SimpleWB(WhiteBalancer):
|
|
56
|
+
# Functions
|
|
57
|
+
def getInputMin(self) -> float: ...
|
|
58
|
+
|
|
59
|
+
def setInputMin(self, val: float) -> None: ...
|
|
60
|
+
|
|
61
|
+
def getInputMax(self) -> float: ...
|
|
62
|
+
|
|
63
|
+
def setInputMax(self, val: float) -> None: ...
|
|
64
|
+
|
|
65
|
+
def getOutputMin(self) -> float: ...
|
|
66
|
+
|
|
67
|
+
def setOutputMin(self, val: float) -> None: ...
|
|
68
|
+
|
|
69
|
+
def getOutputMax(self) -> float: ...
|
|
70
|
+
|
|
71
|
+
def setOutputMax(self, val: float) -> None: ...
|
|
72
|
+
|
|
73
|
+
def getP(self) -> float: ...
|
|
74
|
+
|
|
75
|
+
def setP(self, val: float) -> None: ...
|
|
76
|
+
|
|
77
|
+
|
|
78
|
+
class GrayworldWB(WhiteBalancer):
|
|
79
|
+
# Functions
|
|
80
|
+
def getSaturationThreshold(self) -> float: ...
|
|
81
|
+
|
|
82
|
+
def setSaturationThreshold(self, val: float) -> None: ...
|
|
83
|
+
|
|
84
|
+
|
|
85
|
+
class LearningBasedWB(WhiteBalancer):
|
|
86
|
+
# Functions
|
|
87
|
+
@_typing.overload
|
|
88
|
+
def extractSimpleFeatures(self, src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
89
|
+
@_typing.overload
|
|
90
|
+
def extractSimpleFeatures(self, src: cv2.UMat, dst: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
91
|
+
|
|
92
|
+
def getRangeMaxVal(self) -> int: ...
|
|
93
|
+
|
|
94
|
+
def setRangeMaxVal(self, val: int) -> None: ...
|
|
95
|
+
|
|
96
|
+
def getSaturationThreshold(self) -> float: ...
|
|
97
|
+
|
|
98
|
+
def setSaturationThreshold(self, val: float) -> None: ...
|
|
99
|
+
|
|
100
|
+
def getHistBinNum(self) -> int: ...
|
|
101
|
+
|
|
102
|
+
def setHistBinNum(self, val: int) -> None: ...
|
|
103
|
+
|
|
104
|
+
|
|
105
|
+
|
|
106
|
+
# Functions
|
|
107
|
+
@_typing.overload
|
|
108
|
+
def applyChannelGains(src: cv2.typing.MatLike, gainB: float, gainG: float, gainR: float, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
109
|
+
@_typing.overload
|
|
110
|
+
def applyChannelGains(src: cv2.UMat, gainB: float, gainG: float, gainR: float, dst: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
111
|
+
|
|
112
|
+
@_typing.overload
|
|
113
|
+
def bm3dDenoising(src: cv2.typing.MatLike, dstStep1: cv2.typing.MatLike, dstStep2: cv2.typing.MatLike | None = ..., h: float = ..., templateWindowSize: int = ..., searchWindowSize: int = ..., blockMatchingStep1: int = ..., blockMatchingStep2: int = ..., groupSize: int = ..., slidingStep: int = ..., beta: float = ..., normType: int = ..., step: int = ..., transformType: int = ...) -> tuple[cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
114
|
+
@_typing.overload
|
|
115
|
+
def bm3dDenoising(src: cv2.UMat, dstStep1: cv2.UMat, dstStep2: cv2.UMat | None = ..., h: float = ..., templateWindowSize: int = ..., searchWindowSize: int = ..., blockMatchingStep1: int = ..., blockMatchingStep2: int = ..., groupSize: int = ..., slidingStep: int = ..., beta: float = ..., normType: int = ..., step: int = ..., transformType: int = ...) -> tuple[cv2.UMat, cv2.UMat]: ...
|
|
116
|
+
@_typing.overload
|
|
117
|
+
def bm3dDenoising(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ..., h: float = ..., templateWindowSize: int = ..., searchWindowSize: int = ..., blockMatchingStep1: int = ..., blockMatchingStep2: int = ..., groupSize: int = ..., slidingStep: int = ..., beta: float = ..., normType: int = ..., step: int = ..., transformType: int = ...) -> cv2.typing.MatLike: ...
|
|
118
|
+
@_typing.overload
|
|
119
|
+
def bm3dDenoising(src: cv2.UMat, dst: cv2.UMat | None = ..., h: float = ..., templateWindowSize: int = ..., searchWindowSize: int = ..., blockMatchingStep1: int = ..., blockMatchingStep2: int = ..., groupSize: int = ..., slidingStep: int = ..., beta: float = ..., normType: int = ..., step: int = ..., transformType: int = ...) -> cv2.UMat: ...
|
|
120
|
+
|
|
121
|
+
def createGrayworldWB() -> GrayworldWB: ...
|
|
122
|
+
|
|
123
|
+
def createLearningBasedWB(path_to_model: str = ...) -> LearningBasedWB: ...
|
|
124
|
+
|
|
125
|
+
def createSimpleWB() -> SimpleWB: ...
|
|
126
|
+
|
|
127
|
+
def createTonemapDurand(gamma: float = ..., contrast: float = ..., saturation: float = ..., sigma_color: float = ..., sigma_space: float = ...) -> TonemapDurand: ...
|
|
128
|
+
|
|
129
|
+
def dctDenoising(src: cv2.typing.MatLike, dst: cv2.typing.MatLike, sigma: float, psize: int = ...) -> None: ...
|
|
130
|
+
|
|
131
|
+
def inpaint(src: cv2.typing.MatLike, mask: cv2.typing.MatLike, dst: cv2.typing.MatLike, algorithmType: int) -> None: ...
|
|
132
|
+
|
|
133
|
+
@_typing.overload
|
|
134
|
+
def oilPainting(src: cv2.typing.MatLike, size: int, dynRatio: int, code: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
135
|
+
@_typing.overload
|
|
136
|
+
def oilPainting(src: cv2.UMat, size: int, dynRatio: int, code: int, dst: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
137
|
+
@_typing.overload
|
|
138
|
+
def oilPainting(src: cv2.typing.MatLike, size: int, dynRatio: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
139
|
+
@_typing.overload
|
|
140
|
+
def oilPainting(src: cv2.UMat, size: int, dynRatio: int, dst: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
141
|
+
|
|
142
|
+
|