opencv-contrib-python-headless 4.13.0.90__cp37-abi3-macosx_14_0_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- cv2/.dylibs/libaom.3.13.1.dylib +0 -0
- cv2/.dylibs/libavif.16.3.0.dylib +0 -0
- cv2/.dylibs/libdav1d.7.dylib +0 -0
- cv2/.dylibs/libvmaf.3.dylib +0 -0
- cv2/Error/__init__.pyi +118 -0
- cv2/LICENSE-3RD-PARTY.txt +3513 -0
- cv2/LICENSE.txt +21 -0
- cv2/__init__.py +181 -0
- cv2/__init__.pyi +6858 -0
- cv2/aruco/__init__.pyi +410 -0
- cv2/barcode/__init__.pyi +40 -0
- cv2/bgsegm/__init__.pyi +202 -0
- cv2/bioinspired/__init__.pyi +121 -0
- cv2/ccm/__init__.pyi +167 -0
- cv2/colored_kinfu/__init__.pyi +96 -0
- cv2/config-3.py +24 -0
- cv2/config.py +5 -0
- cv2/cuda/__init__.pyi +553 -0
- cv2/cv2.abi3.so +0 -0
- cv2/data/__init__.py +3 -0
- cv2/data/haarcascade_eye.xml +12213 -0
- cv2/data/haarcascade_eye_tree_eyeglasses.xml +22619 -0
- cv2/data/haarcascade_frontalcatface.xml +14382 -0
- cv2/data/haarcascade_frontalcatface_extended.xml +13394 -0
- cv2/data/haarcascade_frontalface_alt.xml +24350 -0
- cv2/data/haarcascade_frontalface_alt2.xml +20719 -0
- cv2/data/haarcascade_frontalface_alt_tree.xml +96484 -0
- cv2/data/haarcascade_frontalface_default.xml +33314 -0
- cv2/data/haarcascade_fullbody.xml +17030 -0
- cv2/data/haarcascade_lefteye_2splits.xml +7390 -0
- cv2/data/haarcascade_license_plate_rus_16stages.xml +1404 -0
- cv2/data/haarcascade_lowerbody.xml +14056 -0
- cv2/data/haarcascade_profileface.xml +29690 -0
- cv2/data/haarcascade_righteye_2splits.xml +7407 -0
- cv2/data/haarcascade_russian_plate_number.xml +2656 -0
- cv2/data/haarcascade_smile.xml +6729 -0
- cv2/data/haarcascade_upperbody.xml +28134 -0
- cv2/datasets/__init__.pyi +80 -0
- cv2/detail/__init__.pyi +627 -0
- cv2/dnn/__init__.pyi +549 -0
- cv2/dnn_superres/__init__.pyi +37 -0
- cv2/dpm/__init__.pyi +10 -0
- cv2/dynafu/__init__.pyi +43 -0
- cv2/face/__init__.pyi +220 -0
- cv2/fisheye/__init__.pyi +88 -0
- cv2/flann/__init__.pyi +65 -0
- cv2/ft/__init__.pyi +98 -0
- cv2/gapi/__init__.py +323 -0
- cv2/gapi/__init__.pyi +349 -0
- cv2/gapi/core/__init__.pyi +7 -0
- cv2/gapi/core/cpu/__init__.pyi +9 -0
- cv2/gapi/core/fluid/__init__.pyi +9 -0
- cv2/gapi/core/ocl/__init__.pyi +9 -0
- cv2/gapi/ie/__init__.pyi +51 -0
- cv2/gapi/ie/detail/__init__.pyi +12 -0
- cv2/gapi/imgproc/__init__.pyi +5 -0
- cv2/gapi/imgproc/fluid/__init__.pyi +9 -0
- cv2/gapi/oak/__init__.pyi +37 -0
- cv2/gapi/onnx/__init__.pyi +55 -0
- cv2/gapi/onnx/ep/__init__.pyi +63 -0
- cv2/gapi/ot/__init__.pyi +32 -0
- cv2/gapi/ot/cpu/__init__.pyi +9 -0
- cv2/gapi/ov/__init__.pyi +74 -0
- cv2/gapi/own/__init__.pyi +5 -0
- cv2/gapi/own/detail/__init__.pyi +10 -0
- cv2/gapi/render/__init__.pyi +5 -0
- cv2/gapi/render/ocv/__init__.pyi +9 -0
- cv2/gapi/streaming/__init__.pyi +42 -0
- cv2/gapi/video/__init__.pyi +10 -0
- cv2/gapi/wip/__init__.pyi +43 -0
- cv2/gapi/wip/draw/__init__.pyi +119 -0
- cv2/gapi/wip/gst/__init__.pyi +17 -0
- cv2/gapi/wip/onevpl/__init__.pyi +16 -0
- cv2/hfs/__init__.pyi +53 -0
- cv2/img_hash/__init__.pyi +116 -0
- cv2/instr/__init__.pyi +24 -0
- cv2/intensity_transform/__init__.pyi +27 -0
- cv2/ipp/__init__.pyi +14 -0
- cv2/kinfu/__init__.pyi +133 -0
- cv2/kinfu/detail/__init__.pyi +7 -0
- cv2/large_kinfu/__init__.pyi +73 -0
- cv2/legacy/__init__.pyi +93 -0
- cv2/line_descriptor/__init__.pyi +112 -0
- cv2/linemod/__init__.pyi +151 -0
- cv2/load_config_py2.py +6 -0
- cv2/load_config_py3.py +9 -0
- cv2/mat_wrapper/__init__.py +40 -0
- cv2/mcc/__init__.pyi +109 -0
- cv2/misc/__init__.py +1 -0
- cv2/misc/version.py +5 -0
- cv2/ml/__init__.pyi +696 -0
- cv2/motempl/__init__.pyi +29 -0
- cv2/multicalib/__init__.pyi +10 -0
- cv2/ocl/__init__.pyi +252 -0
- cv2/ogl/__init__.pyi +51 -0
- cv2/omnidir/__init__.pyi +68 -0
- cv2/optflow/__init__.pyi +286 -0
- cv2/parallel/__init__.pyi +6 -0
- cv2/phase_unwrapping/__init__.pyi +41 -0
- cv2/plot/__init__.pyi +64 -0
- cv2/ppf_match_3d/__init__.pyi +91 -0
- cv2/py.typed +0 -0
- cv2/quality/__init__.pyi +149 -0
- cv2/rapid/__init__.pyi +91 -0
- cv2/reg/__init__.pyi +210 -0
- cv2/rgbd/__init__.pyi +449 -0
- cv2/saliency/__init__.pyi +117 -0
- cv2/samples/__init__.pyi +12 -0
- cv2/segmentation/__init__.pyi +39 -0
- cv2/signal/__init__.pyi +14 -0
- cv2/stereo/__init__.pyi +88 -0
- cv2/structured_light/__init__.pyi +94 -0
- cv2/text/__init__.pyi +204 -0
- cv2/typing/__init__.py +180 -0
- cv2/utils/__init__.py +14 -0
- cv2/utils/__init__.pyi +110 -0
- cv2/utils/fs/__init__.pyi +6 -0
- cv2/utils/logging/__init__.pyi +22 -0
- cv2/utils/nested/__init__.pyi +31 -0
- cv2/version.py +5 -0
- cv2/videoio_registry/__init__.pyi +31 -0
- cv2/videostab/__init__.pyi +16 -0
- cv2/wechat_qrcode/__init__.pyi +23 -0
- cv2/xfeatures2d/__init__.pyi +537 -0
- cv2/ximgproc/__init__.pyi +746 -0
- cv2/ximgproc/segmentation/__init__.pyi +116 -0
- cv2/xphoto/__init__.pyi +142 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/LICENSE-3RD-PARTY.txt +3513 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/LICENSE.txt +21 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/METADATA +300 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/RECORD +133 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/WHEEL +6 -0
- opencv_contrib_python_headless-4.13.0.90.dist-info/top_level.txt +1 -0
cv2/aruco/__init__.pyi
ADDED
|
@@ -0,0 +1,410 @@
|
|
|
1
|
+
__all__: list[str] = []
|
|
2
|
+
|
|
3
|
+
import cv2
|
|
4
|
+
import cv2.typing
|
|
5
|
+
import typing as _typing
|
|
6
|
+
|
|
7
|
+
|
|
8
|
+
# Enumerations
|
|
9
|
+
CORNER_REFINE_NONE: int
|
|
10
|
+
CORNER_REFINE_SUBPIX: int
|
|
11
|
+
CORNER_REFINE_CONTOUR: int
|
|
12
|
+
CORNER_REFINE_APRILTAG: int
|
|
13
|
+
CornerRefineMethod = int
|
|
14
|
+
"""One of [CORNER_REFINE_NONE, CORNER_REFINE_SUBPIX, CORNER_REFINE_CONTOUR, CORNER_REFINE_APRILTAG]"""
|
|
15
|
+
|
|
16
|
+
DICT_4X4_50: int
|
|
17
|
+
DICT_4X4_100: int
|
|
18
|
+
DICT_4X4_250: int
|
|
19
|
+
DICT_4X4_1000: int
|
|
20
|
+
DICT_5X5_50: int
|
|
21
|
+
DICT_5X5_100: int
|
|
22
|
+
DICT_5X5_250: int
|
|
23
|
+
DICT_5X5_1000: int
|
|
24
|
+
DICT_6X6_50: int
|
|
25
|
+
DICT_6X6_100: int
|
|
26
|
+
DICT_6X6_250: int
|
|
27
|
+
DICT_6X6_1000: int
|
|
28
|
+
DICT_7X7_50: int
|
|
29
|
+
DICT_7X7_100: int
|
|
30
|
+
DICT_7X7_250: int
|
|
31
|
+
DICT_7X7_1000: int
|
|
32
|
+
DICT_ARUCO_ORIGINAL: int
|
|
33
|
+
DICT_APRILTAG_16h5: int
|
|
34
|
+
DICT_APRILTAG_16H5: int
|
|
35
|
+
DICT_APRILTAG_25h9: int
|
|
36
|
+
DICT_APRILTAG_25H9: int
|
|
37
|
+
DICT_APRILTAG_36h10: int
|
|
38
|
+
DICT_APRILTAG_36H10: int
|
|
39
|
+
DICT_APRILTAG_36h11: int
|
|
40
|
+
DICT_APRILTAG_36H11: int
|
|
41
|
+
DICT_ARUCO_MIP_36h12: int
|
|
42
|
+
DICT_ARUCO_MIP_36H12: int
|
|
43
|
+
PredefinedDictionaryType = int
|
|
44
|
+
"""One of [DICT_4X4_50, DICT_4X4_100, DICT_4X4_250, DICT_4X4_1000, DICT_5X5_50, DICT_5X5_100, DICT_5X5_250, DICT_5X5_1000, DICT_6X6_50, DICT_6X6_100, DICT_6X6_250, DICT_6X6_1000, DICT_7X7_50, DICT_7X7_100, DICT_7X7_250, DICT_7X7_1000, DICT_ARUCO_ORIGINAL, DICT_APRILTAG_16h5, DICT_APRILTAG_16H5, DICT_APRILTAG_25h9, DICT_APRILTAG_25H9, DICT_APRILTAG_36h10, DICT_APRILTAG_36H10, DICT_APRILTAG_36h11, DICT_APRILTAG_36H11, DICT_ARUCO_MIP_36h12, DICT_ARUCO_MIP_36H12]"""
|
|
45
|
+
|
|
46
|
+
ARUCO_CCW_CENTER: int
|
|
47
|
+
ARUCO_CW_TOP_LEFT_CORNER: int
|
|
48
|
+
PatternPositionType = int
|
|
49
|
+
"""One of [ARUCO_CCW_CENTER, ARUCO_CW_TOP_LEFT_CORNER]"""
|
|
50
|
+
|
|
51
|
+
|
|
52
|
+
|
|
53
|
+
# Classes
|
|
54
|
+
class Board:
|
|
55
|
+
# Functions
|
|
56
|
+
@_typing.overload
|
|
57
|
+
def __init__(self, objPoints: _typing.Sequence[cv2.typing.MatLike], dictionary: Dictionary, ids: cv2.typing.MatLike) -> None: ...
|
|
58
|
+
@_typing.overload
|
|
59
|
+
def __init__(self, objPoints: _typing.Sequence[cv2.UMat], dictionary: Dictionary, ids: cv2.UMat) -> None: ...
|
|
60
|
+
|
|
61
|
+
def getDictionary(self) -> Dictionary: ...
|
|
62
|
+
|
|
63
|
+
def getObjPoints(self) -> _typing.Sequence[_typing.Sequence[cv2.typing.Point3f]]: ...
|
|
64
|
+
|
|
65
|
+
def getIds(self) -> _typing.Sequence[int]: ...
|
|
66
|
+
|
|
67
|
+
def getRightBottomCorner(self) -> cv2.typing.Point3f: ...
|
|
68
|
+
|
|
69
|
+
@_typing.overload
|
|
70
|
+
def matchImagePoints(self, detectedCorners: _typing.Sequence[cv2.typing.MatLike], detectedIds: cv2.typing.MatLike, objPoints: cv2.typing.MatLike | None = ..., imgPoints: cv2.typing.MatLike | None = ...) -> tuple[cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
71
|
+
@_typing.overload
|
|
72
|
+
def matchImagePoints(self, detectedCorners: _typing.Sequence[cv2.UMat], detectedIds: cv2.UMat, objPoints: cv2.UMat | None = ..., imgPoints: cv2.UMat | None = ...) -> tuple[cv2.UMat, cv2.UMat]: ...
|
|
73
|
+
|
|
74
|
+
@_typing.overload
|
|
75
|
+
def generateImage(self, outSize: cv2.typing.Size, img: cv2.typing.MatLike | None = ..., marginSize: int = ..., borderBits: int = ...) -> cv2.typing.MatLike: ...
|
|
76
|
+
@_typing.overload
|
|
77
|
+
def generateImage(self, outSize: cv2.typing.Size, img: cv2.UMat | None = ..., marginSize: int = ..., borderBits: int = ...) -> cv2.UMat: ...
|
|
78
|
+
|
|
79
|
+
|
|
80
|
+
class GridBoard(Board):
|
|
81
|
+
# Functions
|
|
82
|
+
@_typing.overload
|
|
83
|
+
def __init__(self, size: cv2.typing.Size, markerLength: float, markerSeparation: float, dictionary: Dictionary, ids: cv2.typing.MatLike | None = ...) -> None: ...
|
|
84
|
+
@_typing.overload
|
|
85
|
+
def __init__(self, size: cv2.typing.Size, markerLength: float, markerSeparation: float, dictionary: Dictionary, ids: cv2.UMat | None = ...) -> None: ...
|
|
86
|
+
|
|
87
|
+
def getGridSize(self) -> cv2.typing.Size: ...
|
|
88
|
+
|
|
89
|
+
def getMarkerLength(self) -> float: ...
|
|
90
|
+
|
|
91
|
+
def getMarkerSeparation(self) -> float: ...
|
|
92
|
+
|
|
93
|
+
|
|
94
|
+
class CharucoBoard(Board):
|
|
95
|
+
# Functions
|
|
96
|
+
@_typing.overload
|
|
97
|
+
def __init__(self, size: cv2.typing.Size, squareLength: float, markerLength: float, dictionary: Dictionary, ids: cv2.typing.MatLike | None = ...) -> None: ...
|
|
98
|
+
@_typing.overload
|
|
99
|
+
def __init__(self, size: cv2.typing.Size, squareLength: float, markerLength: float, dictionary: Dictionary, ids: cv2.UMat | None = ...) -> None: ...
|
|
100
|
+
|
|
101
|
+
def setLegacyPattern(self, legacyPattern: bool) -> None: ...
|
|
102
|
+
|
|
103
|
+
def getLegacyPattern(self) -> bool: ...
|
|
104
|
+
|
|
105
|
+
def getChessboardSize(self) -> cv2.typing.Size: ...
|
|
106
|
+
|
|
107
|
+
def getSquareLength(self) -> float: ...
|
|
108
|
+
|
|
109
|
+
def getMarkerLength(self) -> float: ...
|
|
110
|
+
|
|
111
|
+
def getChessboardCorners(self) -> _typing.Sequence[cv2.typing.Point3f]: ...
|
|
112
|
+
|
|
113
|
+
@_typing.overload
|
|
114
|
+
def checkCharucoCornersCollinear(self, charucoIds: cv2.typing.MatLike) -> bool: ...
|
|
115
|
+
@_typing.overload
|
|
116
|
+
def checkCharucoCornersCollinear(self, charucoIds: cv2.UMat) -> bool: ...
|
|
117
|
+
|
|
118
|
+
|
|
119
|
+
class DetectorParameters:
|
|
120
|
+
adaptiveThreshWinSizeMin: int
|
|
121
|
+
adaptiveThreshWinSizeMax: int
|
|
122
|
+
adaptiveThreshWinSizeStep: int
|
|
123
|
+
adaptiveThreshConstant: float
|
|
124
|
+
minMarkerPerimeterRate: float
|
|
125
|
+
maxMarkerPerimeterRate: float
|
|
126
|
+
polygonalApproxAccuracyRate: float
|
|
127
|
+
minCornerDistanceRate: float
|
|
128
|
+
minDistanceToBorder: int
|
|
129
|
+
minMarkerDistanceRate: float
|
|
130
|
+
minGroupDistance: float
|
|
131
|
+
cornerRefinementMethod: int
|
|
132
|
+
cornerRefinementWinSize: int
|
|
133
|
+
relativeCornerRefinmentWinSize: float
|
|
134
|
+
cornerRefinementMaxIterations: int
|
|
135
|
+
cornerRefinementMinAccuracy: float
|
|
136
|
+
markerBorderBits: int
|
|
137
|
+
perspectiveRemovePixelPerCell: int
|
|
138
|
+
perspectiveRemoveIgnoredMarginPerCell: float
|
|
139
|
+
maxErroneousBitsInBorderRate: float
|
|
140
|
+
minOtsuStdDev: float
|
|
141
|
+
errorCorrectionRate: float
|
|
142
|
+
aprilTagQuadDecimate: float
|
|
143
|
+
aprilTagQuadSigma: float
|
|
144
|
+
aprilTagMinClusterPixels: int
|
|
145
|
+
aprilTagMaxNmaxima: int
|
|
146
|
+
aprilTagCriticalRad: float
|
|
147
|
+
aprilTagMaxLineFitMse: float
|
|
148
|
+
aprilTagMinWhiteBlackDiff: int
|
|
149
|
+
aprilTagDeglitch: int
|
|
150
|
+
detectInvertedMarker: bool
|
|
151
|
+
useAruco3Detection: bool
|
|
152
|
+
minSideLengthCanonicalImg: int
|
|
153
|
+
minMarkerLengthRatioOriginalImg: float
|
|
154
|
+
|
|
155
|
+
# Functions
|
|
156
|
+
def __init__(self) -> None: ...
|
|
157
|
+
|
|
158
|
+
def readDetectorParameters(self, fn: cv2.FileNode) -> bool: ...
|
|
159
|
+
|
|
160
|
+
def writeDetectorParameters(self, fs: cv2.FileStorage, name: str = ...) -> bool: ...
|
|
161
|
+
|
|
162
|
+
|
|
163
|
+
class RefineParameters:
|
|
164
|
+
minRepDistance: float
|
|
165
|
+
errorCorrectionRate: float
|
|
166
|
+
checkAllOrders: bool
|
|
167
|
+
|
|
168
|
+
# Functions
|
|
169
|
+
def __init__(self, minRepDistance: float = ..., errorCorrectionRate: float = ..., checkAllOrders: bool = ...) -> None: ...
|
|
170
|
+
|
|
171
|
+
def readRefineParameters(self, fn: cv2.FileNode) -> bool: ...
|
|
172
|
+
|
|
173
|
+
def writeRefineParameters(self, fs: cv2.FileStorage, name: str = ...) -> bool: ...
|
|
174
|
+
|
|
175
|
+
|
|
176
|
+
class ArucoDetector(cv2.Algorithm):
|
|
177
|
+
# Functions
|
|
178
|
+
@_typing.overload
|
|
179
|
+
def __init__(self, dictionary: Dictionary = ..., detectorParams: DetectorParameters = ..., refineParams: RefineParameters = ...) -> None: ...
|
|
180
|
+
@_typing.overload
|
|
181
|
+
def __init__(self, dictionaries: _typing.Sequence[Dictionary], detectorParams: DetectorParameters = ..., refineParams: RefineParameters = ...) -> None: ...
|
|
182
|
+
|
|
183
|
+
@_typing.overload
|
|
184
|
+
def detectMarkers(self, image: cv2.typing.MatLike, corners: _typing.Sequence[cv2.typing.MatLike] | None = ..., ids: cv2.typing.MatLike | None = ..., rejectedImgPoints: _typing.Sequence[cv2.typing.MatLike] | None = ...) -> tuple[_typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike]]: ...
|
|
185
|
+
@_typing.overload
|
|
186
|
+
def detectMarkers(self, image: cv2.UMat, corners: _typing.Sequence[cv2.UMat] | None = ..., ids: cv2.UMat | None = ..., rejectedImgPoints: _typing.Sequence[cv2.UMat] | None = ...) -> tuple[_typing.Sequence[cv2.UMat], cv2.UMat, _typing.Sequence[cv2.UMat]]: ...
|
|
187
|
+
|
|
188
|
+
@_typing.overload
|
|
189
|
+
def detectMarkersWithConfidence(self, image: cv2.typing.MatLike, corners: _typing.Sequence[cv2.typing.MatLike] | None = ..., ids: cv2.typing.MatLike | None = ..., markersConfidence: cv2.typing.MatLike | None = ..., rejectedImgPoints: _typing.Sequence[cv2.typing.MatLike] | None = ...) -> tuple[_typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike]]: ...
|
|
190
|
+
@_typing.overload
|
|
191
|
+
def detectMarkersWithConfidence(self, image: cv2.UMat, corners: _typing.Sequence[cv2.UMat] | None = ..., ids: cv2.UMat | None = ..., markersConfidence: cv2.UMat | None = ..., rejectedImgPoints: _typing.Sequence[cv2.UMat] | None = ...) -> tuple[_typing.Sequence[cv2.UMat], cv2.UMat, cv2.UMat, _typing.Sequence[cv2.UMat]]: ...
|
|
192
|
+
|
|
193
|
+
@_typing.overload
|
|
194
|
+
def refineDetectedMarkers(self, image: cv2.typing.MatLike, board: Board, detectedCorners: _typing.Sequence[cv2.typing.MatLike], detectedIds: cv2.typing.MatLike, rejectedCorners: _typing.Sequence[cv2.typing.MatLike], cameraMatrix: cv2.typing.MatLike | None = ..., distCoeffs: cv2.typing.MatLike | None = ..., recoveredIdxs: cv2.typing.MatLike | None = ...) -> tuple[_typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike]: ...
|
|
195
|
+
@_typing.overload
|
|
196
|
+
def refineDetectedMarkers(self, image: cv2.UMat, board: Board, detectedCorners: _typing.Sequence[cv2.UMat], detectedIds: cv2.UMat, rejectedCorners: _typing.Sequence[cv2.UMat], cameraMatrix: cv2.UMat | None = ..., distCoeffs: cv2.UMat | None = ..., recoveredIdxs: cv2.UMat | None = ...) -> tuple[_typing.Sequence[cv2.UMat], cv2.UMat, _typing.Sequence[cv2.UMat], cv2.UMat]: ...
|
|
197
|
+
|
|
198
|
+
@_typing.overload
|
|
199
|
+
def detectMarkersMultiDict(self, image: cv2.typing.MatLike, corners: _typing.Sequence[cv2.typing.MatLike] | None = ..., ids: cv2.typing.MatLike | None = ..., rejectedImgPoints: _typing.Sequence[cv2.typing.MatLike] | None = ..., dictIndices: cv2.typing.MatLike | None = ...) -> tuple[_typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike]: ...
|
|
200
|
+
@_typing.overload
|
|
201
|
+
def detectMarkersMultiDict(self, image: cv2.UMat, corners: _typing.Sequence[cv2.UMat] | None = ..., ids: cv2.UMat | None = ..., rejectedImgPoints: _typing.Sequence[cv2.UMat] | None = ..., dictIndices: cv2.UMat | None = ...) -> tuple[_typing.Sequence[cv2.UMat], cv2.UMat, _typing.Sequence[cv2.UMat], cv2.UMat]: ...
|
|
202
|
+
|
|
203
|
+
def getDictionary(self) -> Dictionary: ...
|
|
204
|
+
|
|
205
|
+
def setDictionary(self, dictionary: Dictionary) -> None: ...
|
|
206
|
+
|
|
207
|
+
def getDictionaries(self) -> _typing.Sequence[Dictionary]: ...
|
|
208
|
+
|
|
209
|
+
def setDictionaries(self, dictionaries: _typing.Sequence[Dictionary]) -> None: ...
|
|
210
|
+
|
|
211
|
+
def getDetectorParameters(self) -> DetectorParameters: ...
|
|
212
|
+
|
|
213
|
+
def setDetectorParameters(self, detectorParameters: DetectorParameters) -> None: ...
|
|
214
|
+
|
|
215
|
+
def getRefineParameters(self) -> RefineParameters: ...
|
|
216
|
+
|
|
217
|
+
def setRefineParameters(self, refineParameters: RefineParameters) -> None: ...
|
|
218
|
+
|
|
219
|
+
def write(self, fs: cv2.FileStorage, name: str) -> None: ...
|
|
220
|
+
|
|
221
|
+
def read(self, fn: cv2.FileNode) -> None: ...
|
|
222
|
+
|
|
223
|
+
|
|
224
|
+
class Dictionary:
|
|
225
|
+
bytesList: cv2.typing.MatLike
|
|
226
|
+
markerSize: int
|
|
227
|
+
maxCorrectionBits: int
|
|
228
|
+
|
|
229
|
+
# Functions
|
|
230
|
+
@_typing.overload
|
|
231
|
+
def __init__(self) -> None: ...
|
|
232
|
+
@_typing.overload
|
|
233
|
+
def __init__(self, bytesList: cv2.typing.MatLike, _markerSize: int, maxcorr: int = ...) -> None: ...
|
|
234
|
+
|
|
235
|
+
def readDictionary(self, fn: cv2.FileNode) -> bool: ...
|
|
236
|
+
|
|
237
|
+
def writeDictionary(self, fs: cv2.FileStorage, name: str = ...) -> None: ...
|
|
238
|
+
|
|
239
|
+
def identify(self, onlyBits: cv2.typing.MatLike, maxCorrectionRate: float) -> tuple[bool, int, int]: ...
|
|
240
|
+
|
|
241
|
+
@_typing.overload
|
|
242
|
+
def getDistanceToId(self, bits: cv2.typing.MatLike, id: int, allRotations: bool = ...) -> int: ...
|
|
243
|
+
@_typing.overload
|
|
244
|
+
def getDistanceToId(self, bits: cv2.UMat, id: int, allRotations: bool = ...) -> int: ...
|
|
245
|
+
|
|
246
|
+
@_typing.overload
|
|
247
|
+
def generateImageMarker(self, id: int, sidePixels: int, _img: cv2.typing.MatLike | None = ..., borderBits: int = ...) -> cv2.typing.MatLike: ...
|
|
248
|
+
@_typing.overload
|
|
249
|
+
def generateImageMarker(self, id: int, sidePixels: int, _img: cv2.UMat | None = ..., borderBits: int = ...) -> cv2.UMat: ...
|
|
250
|
+
|
|
251
|
+
@staticmethod
|
|
252
|
+
def getByteListFromBits(bits: cv2.typing.MatLike) -> cv2.typing.MatLike: ...
|
|
253
|
+
|
|
254
|
+
@staticmethod
|
|
255
|
+
def getBitsFromByteList(byteList: cv2.typing.MatLike, markerSize: int, rotationId: int = ...) -> cv2.typing.MatLike: ...
|
|
256
|
+
|
|
257
|
+
|
|
258
|
+
class CharucoParameters:
|
|
259
|
+
cameraMatrix: cv2.typing.MatLike
|
|
260
|
+
distCoeffs: cv2.typing.MatLike
|
|
261
|
+
minMarkers: int
|
|
262
|
+
tryRefineMarkers: bool
|
|
263
|
+
checkMarkers: bool
|
|
264
|
+
|
|
265
|
+
# Functions
|
|
266
|
+
def __init__(self) -> None: ...
|
|
267
|
+
|
|
268
|
+
|
|
269
|
+
class CharucoDetector(cv2.Algorithm):
|
|
270
|
+
# Functions
|
|
271
|
+
def __init__(self, board: CharucoBoard, charucoParams: CharucoParameters = ..., detectorParams: DetectorParameters = ..., refineParams: RefineParameters = ...) -> None: ...
|
|
272
|
+
|
|
273
|
+
def getBoard(self) -> CharucoBoard: ...
|
|
274
|
+
|
|
275
|
+
def setBoard(self, board: CharucoBoard) -> None: ...
|
|
276
|
+
|
|
277
|
+
def getCharucoParameters(self) -> CharucoParameters: ...
|
|
278
|
+
|
|
279
|
+
def setCharucoParameters(self, charucoParameters: CharucoParameters) -> None: ...
|
|
280
|
+
|
|
281
|
+
def getDetectorParameters(self) -> DetectorParameters: ...
|
|
282
|
+
|
|
283
|
+
def setDetectorParameters(self, detectorParameters: DetectorParameters) -> None: ...
|
|
284
|
+
|
|
285
|
+
def getRefineParameters(self) -> RefineParameters: ...
|
|
286
|
+
|
|
287
|
+
def setRefineParameters(self, refineParameters: RefineParameters) -> None: ...
|
|
288
|
+
|
|
289
|
+
@_typing.overload
|
|
290
|
+
def detectBoard(self, image: cv2.typing.MatLike, charucoCorners: cv2.typing.MatLike | None = ..., charucoIds: cv2.typing.MatLike | None = ..., markerCorners: _typing.Sequence[cv2.typing.MatLike] | None = ..., markerIds: cv2.typing.MatLike | None = ...) -> tuple[cv2.typing.MatLike, cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike]: ...
|
|
291
|
+
@_typing.overload
|
|
292
|
+
def detectBoard(self, image: cv2.UMat, charucoCorners: cv2.UMat | None = ..., charucoIds: cv2.UMat | None = ..., markerCorners: _typing.Sequence[cv2.UMat] | None = ..., markerIds: cv2.UMat | None = ...) -> tuple[cv2.UMat, cv2.UMat, _typing.Sequence[cv2.UMat], cv2.UMat]: ...
|
|
293
|
+
|
|
294
|
+
@_typing.overload
|
|
295
|
+
def detectDiamonds(self, image: cv2.typing.MatLike, diamondCorners: _typing.Sequence[cv2.typing.MatLike] | None = ..., diamondIds: cv2.typing.MatLike | None = ..., markerCorners: _typing.Sequence[cv2.typing.MatLike] | None = ..., markerIds: cv2.typing.MatLike | None = ...) -> tuple[_typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike]: ...
|
|
296
|
+
@_typing.overload
|
|
297
|
+
def detectDiamonds(self, image: cv2.UMat, diamondCorners: _typing.Sequence[cv2.UMat] | None = ..., diamondIds: cv2.UMat | None = ..., markerCorners: _typing.Sequence[cv2.UMat] | None = ..., markerIds: cv2.UMat | None = ...) -> tuple[_typing.Sequence[cv2.UMat], cv2.UMat, _typing.Sequence[cv2.UMat], cv2.UMat]: ...
|
|
298
|
+
|
|
299
|
+
|
|
300
|
+
class EstimateParameters:
|
|
301
|
+
pattern: PatternPositionType
|
|
302
|
+
useExtrinsicGuess: bool
|
|
303
|
+
solvePnPMethod: int
|
|
304
|
+
|
|
305
|
+
# Functions
|
|
306
|
+
def __init__(self) -> None: ...
|
|
307
|
+
|
|
308
|
+
|
|
309
|
+
|
|
310
|
+
# Functions
|
|
311
|
+
@_typing.overload
|
|
312
|
+
def calibrateCameraAruco(corners: _typing.Sequence[cv2.typing.MatLike], ids: cv2.typing.MatLike, counter: cv2.typing.MatLike, board: Board, imageSize: cv2.typing.Size, cameraMatrix: cv2.typing.MatLike, distCoeffs: cv2.typing.MatLike, rvecs: _typing.Sequence[cv2.typing.MatLike] | None = ..., tvecs: _typing.Sequence[cv2.typing.MatLike] | None = ..., flags: int = ..., criteria: cv2.typing.TermCriteria = ...) -> tuple[float, cv2.typing.MatLike, cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], _typing.Sequence[cv2.typing.MatLike]]: ...
|
|
313
|
+
@_typing.overload
|
|
314
|
+
def calibrateCameraAruco(corners: _typing.Sequence[cv2.UMat], ids: cv2.UMat, counter: cv2.UMat, board: Board, imageSize: cv2.typing.Size, cameraMatrix: cv2.UMat, distCoeffs: cv2.UMat, rvecs: _typing.Sequence[cv2.UMat] | None = ..., tvecs: _typing.Sequence[cv2.UMat] | None = ..., flags: int = ..., criteria: cv2.typing.TermCriteria = ...) -> tuple[float, cv2.UMat, cv2.UMat, _typing.Sequence[cv2.UMat], _typing.Sequence[cv2.UMat]]: ...
|
|
315
|
+
|
|
316
|
+
@_typing.overload
|
|
317
|
+
def calibrateCameraArucoExtended(corners: _typing.Sequence[cv2.typing.MatLike], ids: cv2.typing.MatLike, counter: cv2.typing.MatLike, board: Board, imageSize: cv2.typing.Size, cameraMatrix: cv2.typing.MatLike, distCoeffs: cv2.typing.MatLike, rvecs: _typing.Sequence[cv2.typing.MatLike] | None = ..., tvecs: _typing.Sequence[cv2.typing.MatLike] | None = ..., stdDeviationsIntrinsics: cv2.typing.MatLike | None = ..., stdDeviationsExtrinsics: cv2.typing.MatLike | None = ..., perViewErrors: cv2.typing.MatLike | None = ..., flags: int = ..., criteria: cv2.typing.TermCriteria = ...) -> tuple[float, cv2.typing.MatLike, cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], _typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
318
|
+
@_typing.overload
|
|
319
|
+
def calibrateCameraArucoExtended(corners: _typing.Sequence[cv2.UMat], ids: cv2.UMat, counter: cv2.UMat, board: Board, imageSize: cv2.typing.Size, cameraMatrix: cv2.UMat, distCoeffs: cv2.UMat, rvecs: _typing.Sequence[cv2.UMat] | None = ..., tvecs: _typing.Sequence[cv2.UMat] | None = ..., stdDeviationsIntrinsics: cv2.UMat | None = ..., stdDeviationsExtrinsics: cv2.UMat | None = ..., perViewErrors: cv2.UMat | None = ..., flags: int = ..., criteria: cv2.typing.TermCriteria = ...) -> tuple[float, cv2.UMat, cv2.UMat, _typing.Sequence[cv2.UMat], _typing.Sequence[cv2.UMat], cv2.UMat, cv2.UMat, cv2.UMat]: ...
|
|
320
|
+
|
|
321
|
+
@_typing.overload
|
|
322
|
+
def calibrateCameraCharuco(charucoCorners: _typing.Sequence[cv2.typing.MatLike], charucoIds: _typing.Sequence[cv2.typing.MatLike], board: CharucoBoard, imageSize: cv2.typing.Size, cameraMatrix: cv2.typing.MatLike, distCoeffs: cv2.typing.MatLike, rvecs: _typing.Sequence[cv2.typing.MatLike] | None = ..., tvecs: _typing.Sequence[cv2.typing.MatLike] | None = ..., flags: int = ..., criteria: cv2.typing.TermCriteria = ...) -> tuple[float, cv2.typing.MatLike, cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], _typing.Sequence[cv2.typing.MatLike]]: ...
|
|
323
|
+
@_typing.overload
|
|
324
|
+
def calibrateCameraCharuco(charucoCorners: _typing.Sequence[cv2.UMat], charucoIds: _typing.Sequence[cv2.UMat], board: CharucoBoard, imageSize: cv2.typing.Size, cameraMatrix: cv2.UMat, distCoeffs: cv2.UMat, rvecs: _typing.Sequence[cv2.UMat] | None = ..., tvecs: _typing.Sequence[cv2.UMat] | None = ..., flags: int = ..., criteria: cv2.typing.TermCriteria = ...) -> tuple[float, cv2.UMat, cv2.UMat, _typing.Sequence[cv2.UMat], _typing.Sequence[cv2.UMat]]: ...
|
|
325
|
+
|
|
326
|
+
@_typing.overload
|
|
327
|
+
def calibrateCameraCharucoExtended(charucoCorners: _typing.Sequence[cv2.typing.MatLike], charucoIds: _typing.Sequence[cv2.typing.MatLike], board: CharucoBoard, imageSize: cv2.typing.Size, cameraMatrix: cv2.typing.MatLike, distCoeffs: cv2.typing.MatLike, rvecs: _typing.Sequence[cv2.typing.MatLike] | None = ..., tvecs: _typing.Sequence[cv2.typing.MatLike] | None = ..., stdDeviationsIntrinsics: cv2.typing.MatLike | None = ..., stdDeviationsExtrinsics: cv2.typing.MatLike | None = ..., perViewErrors: cv2.typing.MatLike | None = ..., flags: int = ..., criteria: cv2.typing.TermCriteria = ...) -> tuple[float, cv2.typing.MatLike, cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], _typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
328
|
+
@_typing.overload
|
|
329
|
+
def calibrateCameraCharucoExtended(charucoCorners: _typing.Sequence[cv2.UMat], charucoIds: _typing.Sequence[cv2.UMat], board: CharucoBoard, imageSize: cv2.typing.Size, cameraMatrix: cv2.UMat, distCoeffs: cv2.UMat, rvecs: _typing.Sequence[cv2.UMat] | None = ..., tvecs: _typing.Sequence[cv2.UMat] | None = ..., stdDeviationsIntrinsics: cv2.UMat | None = ..., stdDeviationsExtrinsics: cv2.UMat | None = ..., perViewErrors: cv2.UMat | None = ..., flags: int = ..., criteria: cv2.typing.TermCriteria = ...) -> tuple[float, cv2.UMat, cv2.UMat, _typing.Sequence[cv2.UMat], _typing.Sequence[cv2.UMat], cv2.UMat, cv2.UMat, cv2.UMat]: ...
|
|
330
|
+
|
|
331
|
+
@_typing.overload
|
|
332
|
+
def detectCharucoDiamond(image: cv2.typing.MatLike, markerCorners: _typing.Sequence[cv2.typing.MatLike], markerIds: cv2.typing.MatLike, squareMarkerLengthRate: float, diamondCorners: _typing.Sequence[cv2.typing.MatLike] | None = ..., diamondIds: cv2.typing.MatLike | None = ..., cameraMatrix: cv2.typing.MatLike | None = ..., distCoeffs: cv2.typing.MatLike | None = ..., dictionary: Dictionary = ...) -> tuple[_typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike]: ...
|
|
333
|
+
@_typing.overload
|
|
334
|
+
def detectCharucoDiamond(image: cv2.UMat, markerCorners: _typing.Sequence[cv2.UMat], markerIds: cv2.UMat, squareMarkerLengthRate: float, diamondCorners: _typing.Sequence[cv2.UMat] | None = ..., diamondIds: cv2.UMat | None = ..., cameraMatrix: cv2.UMat | None = ..., distCoeffs: cv2.UMat | None = ..., dictionary: Dictionary = ...) -> tuple[_typing.Sequence[cv2.UMat], cv2.UMat]: ...
|
|
335
|
+
|
|
336
|
+
@_typing.overload
|
|
337
|
+
def detectMarkers(image: cv2.typing.MatLike, dictionary: Dictionary, corners: _typing.Sequence[cv2.typing.MatLike] | None = ..., ids: cv2.typing.MatLike | None = ..., parameters: DetectorParameters = ..., rejectedImgPoints: _typing.Sequence[cv2.typing.MatLike] | None = ...) -> tuple[_typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike]]: ...
|
|
338
|
+
@_typing.overload
|
|
339
|
+
def detectMarkers(image: cv2.UMat, dictionary: Dictionary, corners: _typing.Sequence[cv2.UMat] | None = ..., ids: cv2.UMat | None = ..., parameters: DetectorParameters = ..., rejectedImgPoints: _typing.Sequence[cv2.UMat] | None = ...) -> tuple[_typing.Sequence[cv2.UMat], cv2.UMat, _typing.Sequence[cv2.UMat]]: ...
|
|
340
|
+
|
|
341
|
+
@_typing.overload
|
|
342
|
+
def drawCharucoDiamond(dictionary: Dictionary, ids: cv2.typing.Vec4i, squareLength: int, markerLength: int, img: cv2.typing.MatLike | None = ..., marginSize: int = ..., borderBits: int = ...) -> cv2.typing.MatLike: ...
|
|
343
|
+
@_typing.overload
|
|
344
|
+
def drawCharucoDiamond(dictionary: Dictionary, ids: cv2.typing.Vec4i, squareLength: int, markerLength: int, img: cv2.UMat | None = ..., marginSize: int = ..., borderBits: int = ...) -> cv2.UMat: ...
|
|
345
|
+
|
|
346
|
+
@_typing.overload
|
|
347
|
+
def drawDetectedCornersCharuco(image: cv2.typing.MatLike, charucoCorners: cv2.typing.MatLike, charucoIds: cv2.typing.MatLike | None = ..., cornerColor: cv2.typing.Scalar = ...) -> cv2.typing.MatLike: ...
|
|
348
|
+
@_typing.overload
|
|
349
|
+
def drawDetectedCornersCharuco(image: cv2.UMat, charucoCorners: cv2.UMat, charucoIds: cv2.UMat | None = ..., cornerColor: cv2.typing.Scalar = ...) -> cv2.UMat: ...
|
|
350
|
+
|
|
351
|
+
@_typing.overload
|
|
352
|
+
def drawDetectedDiamonds(image: cv2.typing.MatLike, diamondCorners: _typing.Sequence[cv2.typing.MatLike], diamondIds: cv2.typing.MatLike | None = ..., borderColor: cv2.typing.Scalar = ...) -> cv2.typing.MatLike: ...
|
|
353
|
+
@_typing.overload
|
|
354
|
+
def drawDetectedDiamonds(image: cv2.UMat, diamondCorners: _typing.Sequence[cv2.UMat], diamondIds: cv2.UMat | None = ..., borderColor: cv2.typing.Scalar = ...) -> cv2.UMat: ...
|
|
355
|
+
|
|
356
|
+
@_typing.overload
|
|
357
|
+
def drawDetectedMarkers(image: cv2.typing.MatLike, corners: _typing.Sequence[cv2.typing.MatLike], ids: cv2.typing.MatLike | None = ..., borderColor: cv2.typing.Scalar = ...) -> cv2.typing.MatLike: ...
|
|
358
|
+
@_typing.overload
|
|
359
|
+
def drawDetectedMarkers(image: cv2.UMat, corners: _typing.Sequence[cv2.UMat], ids: cv2.UMat | None = ..., borderColor: cv2.typing.Scalar = ...) -> cv2.UMat: ...
|
|
360
|
+
|
|
361
|
+
@_typing.overload
|
|
362
|
+
def drawPlanarBoard(board: Board, outSize: cv2.typing.Size, marginSize: int, borderBits: int, img: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
363
|
+
@_typing.overload
|
|
364
|
+
def drawPlanarBoard(board: Board, outSize: cv2.typing.Size, marginSize: int, borderBits: int, img: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
365
|
+
|
|
366
|
+
@_typing.overload
|
|
367
|
+
def estimatePoseBoard(corners: _typing.Sequence[cv2.typing.MatLike], ids: cv2.typing.MatLike, board: Board, cameraMatrix: cv2.typing.MatLike, distCoeffs: cv2.typing.MatLike, rvec: cv2.typing.MatLike, tvec: cv2.typing.MatLike, useExtrinsicGuess: bool = ...) -> tuple[int, cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
368
|
+
@_typing.overload
|
|
369
|
+
def estimatePoseBoard(corners: _typing.Sequence[cv2.UMat], ids: cv2.UMat, board: Board, cameraMatrix: cv2.UMat, distCoeffs: cv2.UMat, rvec: cv2.UMat, tvec: cv2.UMat, useExtrinsicGuess: bool = ...) -> tuple[int, cv2.UMat, cv2.UMat]: ...
|
|
370
|
+
|
|
371
|
+
@_typing.overload
|
|
372
|
+
def estimatePoseCharucoBoard(charucoCorners: cv2.typing.MatLike, charucoIds: cv2.typing.MatLike, board: CharucoBoard, cameraMatrix: cv2.typing.MatLike, distCoeffs: cv2.typing.MatLike, rvec: cv2.typing.MatLike, tvec: cv2.typing.MatLike, useExtrinsicGuess: bool = ...) -> tuple[bool, cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
373
|
+
@_typing.overload
|
|
374
|
+
def estimatePoseCharucoBoard(charucoCorners: cv2.UMat, charucoIds: cv2.UMat, board: CharucoBoard, cameraMatrix: cv2.UMat, distCoeffs: cv2.UMat, rvec: cv2.UMat, tvec: cv2.UMat, useExtrinsicGuess: bool = ...) -> tuple[bool, cv2.UMat, cv2.UMat]: ...
|
|
375
|
+
|
|
376
|
+
@_typing.overload
|
|
377
|
+
def estimatePoseSingleMarkers(corners: _typing.Sequence[cv2.typing.MatLike], markerLength: float, cameraMatrix: cv2.typing.MatLike, distCoeffs: cv2.typing.MatLike, rvecs: cv2.typing.MatLike | None = ..., tvecs: cv2.typing.MatLike | None = ..., objPoints: cv2.typing.MatLike | None = ..., estimateParameters: EstimateParameters = ...) -> tuple[cv2.typing.MatLike, cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
378
|
+
@_typing.overload
|
|
379
|
+
def estimatePoseSingleMarkers(corners: _typing.Sequence[cv2.UMat], markerLength: float, cameraMatrix: cv2.UMat, distCoeffs: cv2.UMat, rvecs: cv2.UMat | None = ..., tvecs: cv2.UMat | None = ..., objPoints: cv2.UMat | None = ..., estimateParameters: EstimateParameters = ...) -> tuple[cv2.UMat, cv2.UMat, cv2.UMat]: ...
|
|
380
|
+
|
|
381
|
+
def extendDictionary(nMarkers: int, markerSize: int, baseDictionary: Dictionary = ..., randomSeed: int = ...) -> Dictionary: ...
|
|
382
|
+
|
|
383
|
+
@_typing.overload
|
|
384
|
+
def generateImageMarker(dictionary: Dictionary, id: int, sidePixels: int, img: cv2.typing.MatLike | None = ..., borderBits: int = ...) -> cv2.typing.MatLike: ...
|
|
385
|
+
@_typing.overload
|
|
386
|
+
def generateImageMarker(dictionary: Dictionary, id: int, sidePixels: int, img: cv2.UMat | None = ..., borderBits: int = ...) -> cv2.UMat: ...
|
|
387
|
+
|
|
388
|
+
@_typing.overload
|
|
389
|
+
def getBoardObjectAndImagePoints(board: Board, detectedCorners: _typing.Sequence[cv2.typing.MatLike], detectedIds: cv2.typing.MatLike, objPoints: cv2.typing.MatLike | None = ..., imgPoints: cv2.typing.MatLike | None = ...) -> tuple[cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
390
|
+
@_typing.overload
|
|
391
|
+
def getBoardObjectAndImagePoints(board: Board, detectedCorners: _typing.Sequence[cv2.UMat], detectedIds: cv2.UMat, objPoints: cv2.UMat | None = ..., imgPoints: cv2.UMat | None = ...) -> tuple[cv2.UMat, cv2.UMat]: ...
|
|
392
|
+
|
|
393
|
+
def getPredefinedDictionary(dict: int) -> Dictionary: ...
|
|
394
|
+
|
|
395
|
+
@_typing.overload
|
|
396
|
+
def interpolateCornersCharuco(markerCorners: _typing.Sequence[cv2.typing.MatLike], markerIds: cv2.typing.MatLike, image: cv2.typing.MatLike, board: CharucoBoard, charucoCorners: cv2.typing.MatLike | None = ..., charucoIds: cv2.typing.MatLike | None = ..., cameraMatrix: cv2.typing.MatLike | None = ..., distCoeffs: cv2.typing.MatLike | None = ..., minMarkers: int = ...) -> tuple[int, cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
397
|
+
@_typing.overload
|
|
398
|
+
def interpolateCornersCharuco(markerCorners: _typing.Sequence[cv2.UMat], markerIds: cv2.UMat, image: cv2.UMat, board: CharucoBoard, charucoCorners: cv2.UMat | None = ..., charucoIds: cv2.UMat | None = ..., cameraMatrix: cv2.UMat | None = ..., distCoeffs: cv2.UMat | None = ..., minMarkers: int = ...) -> tuple[int, cv2.UMat, cv2.UMat]: ...
|
|
399
|
+
|
|
400
|
+
@_typing.overload
|
|
401
|
+
def refineDetectedMarkers(image: cv2.typing.MatLike, board: Board, detectedCorners: _typing.Sequence[cv2.typing.MatLike], detectedIds: cv2.typing.MatLike, rejectedCorners: _typing.Sequence[cv2.typing.MatLike], cameraMatrix: cv2.typing.MatLike | None = ..., distCoeffs: cv2.typing.MatLike | None = ..., minRepDistance: float = ..., errorCorrectionRate: float = ..., checkAllOrders: bool = ..., recoveredIdxs: cv2.typing.MatLike | None = ..., parameters: DetectorParameters = ...) -> tuple[_typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike, _typing.Sequence[cv2.typing.MatLike], cv2.typing.MatLike]: ...
|
|
402
|
+
@_typing.overload
|
|
403
|
+
def refineDetectedMarkers(image: cv2.UMat, board: Board, detectedCorners: _typing.Sequence[cv2.UMat], detectedIds: cv2.UMat, rejectedCorners: _typing.Sequence[cv2.UMat], cameraMatrix: cv2.UMat | None = ..., distCoeffs: cv2.UMat | None = ..., minRepDistance: float = ..., errorCorrectionRate: float = ..., checkAllOrders: bool = ..., recoveredIdxs: cv2.UMat | None = ..., parameters: DetectorParameters = ...) -> tuple[_typing.Sequence[cv2.UMat], cv2.UMat, _typing.Sequence[cv2.UMat], cv2.UMat]: ...
|
|
404
|
+
|
|
405
|
+
@_typing.overload
|
|
406
|
+
def testCharucoCornersCollinear(board: CharucoBoard, charucoIds: cv2.typing.MatLike) -> bool: ...
|
|
407
|
+
@_typing.overload
|
|
408
|
+
def testCharucoCornersCollinear(board: CharucoBoard, charucoIds: cv2.UMat) -> bool: ...
|
|
409
|
+
|
|
410
|
+
|
cv2/barcode/__init__.pyi
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
1
|
+
__all__: list[str] = []
|
|
2
|
+
|
|
3
|
+
import cv2
|
|
4
|
+
import cv2.typing
|
|
5
|
+
import os
|
|
6
|
+
import typing as _typing
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
# Classes
|
|
10
|
+
class BarcodeDetector(cv2.GraphicalCodeDetector):
|
|
11
|
+
# Functions
|
|
12
|
+
@_typing.overload
|
|
13
|
+
def __init__(self) -> None: ...
|
|
14
|
+
@_typing.overload
|
|
15
|
+
def __init__(self, prototxt_path: str | os.PathLike[str], model_path: str | os.PathLike[str]) -> None: ...
|
|
16
|
+
|
|
17
|
+
@_typing.overload
|
|
18
|
+
def decodeWithType(self, img: cv2.typing.MatLike, points: cv2.typing.MatLike) -> tuple[bool, _typing.Sequence[str], _typing.Sequence[str]]: ...
|
|
19
|
+
@_typing.overload
|
|
20
|
+
def decodeWithType(self, img: cv2.UMat, points: cv2.UMat) -> tuple[bool, _typing.Sequence[str], _typing.Sequence[str]]: ...
|
|
21
|
+
|
|
22
|
+
@_typing.overload
|
|
23
|
+
def detectAndDecodeWithType(self, img: cv2.typing.MatLike, points: cv2.typing.MatLike | None = ...) -> tuple[bool, _typing.Sequence[str], _typing.Sequence[str], cv2.typing.MatLike]: ...
|
|
24
|
+
@_typing.overload
|
|
25
|
+
def detectAndDecodeWithType(self, img: cv2.UMat, points: cv2.UMat | None = ...) -> tuple[bool, _typing.Sequence[str], _typing.Sequence[str], cv2.UMat]: ...
|
|
26
|
+
|
|
27
|
+
def getDownsamplingThreshold(self) -> float: ...
|
|
28
|
+
|
|
29
|
+
def setDownsamplingThreshold(self, thresh: float) -> BarcodeDetector: ...
|
|
30
|
+
|
|
31
|
+
def getDetectorScales(self) -> _typing.Sequence[float]: ...
|
|
32
|
+
|
|
33
|
+
def setDetectorScales(self, sizes: _typing.Sequence[float]) -> BarcodeDetector: ...
|
|
34
|
+
|
|
35
|
+
def getGradientThreshold(self) -> float: ...
|
|
36
|
+
|
|
37
|
+
def setGradientThreshold(self, thresh: float) -> BarcodeDetector: ...
|
|
38
|
+
|
|
39
|
+
|
|
40
|
+
|
cv2/bgsegm/__init__.pyi
ADDED
|
@@ -0,0 +1,202 @@
|
|
|
1
|
+
__all__: list[str] = []
|
|
2
|
+
|
|
3
|
+
import cv2
|
|
4
|
+
import cv2.typing
|
|
5
|
+
import typing as _typing
|
|
6
|
+
|
|
7
|
+
|
|
8
|
+
# Enumerations
|
|
9
|
+
LSBP_CAMERA_MOTION_COMPENSATION_NONE: int
|
|
10
|
+
LSBP_CAMERA_MOTION_COMPENSATION_LK: int
|
|
11
|
+
LSBPCameraMotionCompensation = int
|
|
12
|
+
"""One of [LSBP_CAMERA_MOTION_COMPENSATION_NONE, LSBP_CAMERA_MOTION_COMPENSATION_LK]"""
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
|
|
16
|
+
# Classes
|
|
17
|
+
class BackgroundSubtractorMOG(cv2.BackgroundSubtractor):
|
|
18
|
+
# Functions
|
|
19
|
+
@_typing.overload
|
|
20
|
+
def apply(self, image: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
21
|
+
@_typing.overload
|
|
22
|
+
def apply(self, image: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
23
|
+
@_typing.overload
|
|
24
|
+
def apply(self, image: cv2.typing.MatLike, knownForegroundMask: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
25
|
+
@_typing.overload
|
|
26
|
+
def apply(self, image: cv2.UMat, knownForegroundMask: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
27
|
+
|
|
28
|
+
def getHistory(self) -> int: ...
|
|
29
|
+
|
|
30
|
+
def setHistory(self, nframes: int) -> None: ...
|
|
31
|
+
|
|
32
|
+
def getNMixtures(self) -> int: ...
|
|
33
|
+
|
|
34
|
+
def setNMixtures(self, nmix: int) -> None: ...
|
|
35
|
+
|
|
36
|
+
def getBackgroundRatio(self) -> float: ...
|
|
37
|
+
|
|
38
|
+
def setBackgroundRatio(self, backgroundRatio: float) -> None: ...
|
|
39
|
+
|
|
40
|
+
def getNoiseSigma(self) -> float: ...
|
|
41
|
+
|
|
42
|
+
def setNoiseSigma(self, noiseSigma: float) -> None: ...
|
|
43
|
+
|
|
44
|
+
|
|
45
|
+
class BackgroundSubtractorGMG(cv2.BackgroundSubtractor):
|
|
46
|
+
# Functions
|
|
47
|
+
@_typing.overload
|
|
48
|
+
def apply(self, image: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
49
|
+
@_typing.overload
|
|
50
|
+
def apply(self, image: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
51
|
+
@_typing.overload
|
|
52
|
+
def apply(self, image: cv2.typing.MatLike, knownForegroundMask: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
53
|
+
@_typing.overload
|
|
54
|
+
def apply(self, image: cv2.UMat, knownForegroundMask: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
55
|
+
|
|
56
|
+
@_typing.overload
|
|
57
|
+
def getBackgroundImage(self, backgroundImage: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
58
|
+
@_typing.overload
|
|
59
|
+
def getBackgroundImage(self, backgroundImage: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
60
|
+
|
|
61
|
+
def getMaxFeatures(self) -> int: ...
|
|
62
|
+
|
|
63
|
+
def setMaxFeatures(self, maxFeatures: int) -> None: ...
|
|
64
|
+
|
|
65
|
+
def getDefaultLearningRate(self) -> float: ...
|
|
66
|
+
|
|
67
|
+
def setDefaultLearningRate(self, lr: float) -> None: ...
|
|
68
|
+
|
|
69
|
+
def getNumFrames(self) -> int: ...
|
|
70
|
+
|
|
71
|
+
def setNumFrames(self, nframes: int) -> None: ...
|
|
72
|
+
|
|
73
|
+
def getQuantizationLevels(self) -> int: ...
|
|
74
|
+
|
|
75
|
+
def setQuantizationLevels(self, nlevels: int) -> None: ...
|
|
76
|
+
|
|
77
|
+
def getBackgroundPrior(self) -> float: ...
|
|
78
|
+
|
|
79
|
+
def setBackgroundPrior(self, bgprior: float) -> None: ...
|
|
80
|
+
|
|
81
|
+
def getSmoothingRadius(self) -> int: ...
|
|
82
|
+
|
|
83
|
+
def setSmoothingRadius(self, radius: int) -> None: ...
|
|
84
|
+
|
|
85
|
+
def getDecisionThreshold(self) -> float: ...
|
|
86
|
+
|
|
87
|
+
def setDecisionThreshold(self, thresh: float) -> None: ...
|
|
88
|
+
|
|
89
|
+
def getUpdateBackgroundModel(self) -> bool: ...
|
|
90
|
+
|
|
91
|
+
def setUpdateBackgroundModel(self, update: bool) -> None: ...
|
|
92
|
+
|
|
93
|
+
def getMinVal(self) -> float: ...
|
|
94
|
+
|
|
95
|
+
def setMinVal(self, val: float) -> None: ...
|
|
96
|
+
|
|
97
|
+
def getMaxVal(self) -> float: ...
|
|
98
|
+
|
|
99
|
+
def setMaxVal(self, val: float) -> None: ...
|
|
100
|
+
|
|
101
|
+
|
|
102
|
+
class BackgroundSubtractorCNT(cv2.BackgroundSubtractor):
|
|
103
|
+
# Functions
|
|
104
|
+
@_typing.overload
|
|
105
|
+
def apply(self, image: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
106
|
+
@_typing.overload
|
|
107
|
+
def apply(self, image: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
108
|
+
@_typing.overload
|
|
109
|
+
def apply(self, image: cv2.typing.MatLike, knownForegroundMask: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
110
|
+
@_typing.overload
|
|
111
|
+
def apply(self, image: cv2.UMat, knownForegroundMask: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
112
|
+
|
|
113
|
+
@_typing.overload
|
|
114
|
+
def getBackgroundImage(self, backgroundImage: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
115
|
+
@_typing.overload
|
|
116
|
+
def getBackgroundImage(self, backgroundImage: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
117
|
+
|
|
118
|
+
def getMinPixelStability(self) -> int: ...
|
|
119
|
+
|
|
120
|
+
def setMinPixelStability(self, value: int) -> None: ...
|
|
121
|
+
|
|
122
|
+
def getMaxPixelStability(self) -> int: ...
|
|
123
|
+
|
|
124
|
+
def setMaxPixelStability(self, value: int) -> None: ...
|
|
125
|
+
|
|
126
|
+
def getUseHistory(self) -> bool: ...
|
|
127
|
+
|
|
128
|
+
def setUseHistory(self, value: bool) -> None: ...
|
|
129
|
+
|
|
130
|
+
def getIsParallel(self) -> bool: ...
|
|
131
|
+
|
|
132
|
+
def setIsParallel(self, value: bool) -> None: ...
|
|
133
|
+
|
|
134
|
+
|
|
135
|
+
class BackgroundSubtractorGSOC(cv2.BackgroundSubtractor):
|
|
136
|
+
# Functions
|
|
137
|
+
@_typing.overload
|
|
138
|
+
def apply(self, image: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
139
|
+
@_typing.overload
|
|
140
|
+
def apply(self, image: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
141
|
+
@_typing.overload
|
|
142
|
+
def apply(self, image: cv2.typing.MatLike, knownForegroundMask: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
143
|
+
@_typing.overload
|
|
144
|
+
def apply(self, image: cv2.UMat, knownForegroundMask: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
145
|
+
|
|
146
|
+
@_typing.overload
|
|
147
|
+
def getBackgroundImage(self, backgroundImage: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
148
|
+
@_typing.overload
|
|
149
|
+
def getBackgroundImage(self, backgroundImage: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
150
|
+
|
|
151
|
+
|
|
152
|
+
class BackgroundSubtractorLSBP(cv2.BackgroundSubtractor):
|
|
153
|
+
# Functions
|
|
154
|
+
@_typing.overload
|
|
155
|
+
def apply(self, image: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
156
|
+
@_typing.overload
|
|
157
|
+
def apply(self, image: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
158
|
+
@_typing.overload
|
|
159
|
+
def apply(self, image: cv2.typing.MatLike, knownForegroundMask: cv2.typing.MatLike, fgmask: cv2.typing.MatLike | None = ..., learningRate: float = ...) -> cv2.typing.MatLike: ...
|
|
160
|
+
@_typing.overload
|
|
161
|
+
def apply(self, image: cv2.UMat, knownForegroundMask: cv2.UMat, fgmask: cv2.UMat | None = ..., learningRate: float = ...) -> cv2.UMat: ...
|
|
162
|
+
|
|
163
|
+
@_typing.overload
|
|
164
|
+
def getBackgroundImage(self, backgroundImage: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ...
|
|
165
|
+
@_typing.overload
|
|
166
|
+
def getBackgroundImage(self, backgroundImage: cv2.UMat | None = ...) -> cv2.UMat: ...
|
|
167
|
+
|
|
168
|
+
|
|
169
|
+
class BackgroundSubtractorLSBPDesc:
|
|
170
|
+
...
|
|
171
|
+
|
|
172
|
+
class SyntheticSequenceGenerator(cv2.Algorithm):
|
|
173
|
+
# Functions
|
|
174
|
+
@_typing.overload
|
|
175
|
+
def __init__(self, background: cv2.typing.MatLike, object: cv2.typing.MatLike, amplitude: float, wavelength: float, wavespeed: float, objspeed: float) -> None: ...
|
|
176
|
+
@_typing.overload
|
|
177
|
+
def __init__(self, background: cv2.UMat, object: cv2.UMat, amplitude: float, wavelength: float, wavespeed: float, objspeed: float) -> None: ...
|
|
178
|
+
|
|
179
|
+
@_typing.overload
|
|
180
|
+
def getNextFrame(self, frame: cv2.typing.MatLike | None = ..., gtMask: cv2.typing.MatLike | None = ...) -> tuple[cv2.typing.MatLike, cv2.typing.MatLike]: ...
|
|
181
|
+
@_typing.overload
|
|
182
|
+
def getNextFrame(self, frame: cv2.UMat | None = ..., gtMask: cv2.UMat | None = ...) -> tuple[cv2.UMat, cv2.UMat]: ...
|
|
183
|
+
|
|
184
|
+
|
|
185
|
+
|
|
186
|
+
# Functions
|
|
187
|
+
def createBackgroundSubtractorCNT(minPixelStability: int = ..., useHistory: bool = ..., maxPixelStability: int = ..., isParallel: bool = ...) -> BackgroundSubtractorCNT: ...
|
|
188
|
+
|
|
189
|
+
def createBackgroundSubtractorGMG(initializationFrames: int = ..., decisionThreshold: float = ...) -> BackgroundSubtractorGMG: ...
|
|
190
|
+
|
|
191
|
+
def createBackgroundSubtractorGSOC(mc: int = ..., nSamples: int = ..., replaceRate: float = ..., propagationRate: float = ..., hitsThreshold: int = ..., alpha: float = ..., beta: float = ..., blinkingSupressionDecay: float = ..., blinkingSupressionMultiplier: float = ..., noiseRemovalThresholdFacBG: float = ..., noiseRemovalThresholdFacFG: float = ...) -> BackgroundSubtractorGSOC: ...
|
|
192
|
+
|
|
193
|
+
def createBackgroundSubtractorLSBP(mc: int = ..., nSamples: int = ..., LSBPRadius: int = ..., Tlower: float = ..., Tupper: float = ..., Tinc: float = ..., Tdec: float = ..., Rscale: float = ..., Rincdec: float = ..., noiseRemovalThresholdFacBG: float = ..., noiseRemovalThresholdFacFG: float = ..., LSBPthreshold: int = ..., minCount: int = ...) -> BackgroundSubtractorLSBP: ...
|
|
194
|
+
|
|
195
|
+
def createBackgroundSubtractorMOG(history: int = ..., nmixtures: int = ..., backgroundRatio: float = ..., noiseSigma: float = ...) -> BackgroundSubtractorMOG: ...
|
|
196
|
+
|
|
197
|
+
@_typing.overload
|
|
198
|
+
def createSyntheticSequenceGenerator(background: cv2.typing.MatLike, object: cv2.typing.MatLike, amplitude: float = ..., wavelength: float = ..., wavespeed: float = ..., objspeed: float = ...) -> SyntheticSequenceGenerator: ...
|
|
199
|
+
@_typing.overload
|
|
200
|
+
def createSyntheticSequenceGenerator(background: cv2.UMat, object: cv2.UMat, amplitude: float = ..., wavelength: float = ..., wavespeed: float = ..., objspeed: float = ...) -> SyntheticSequenceGenerator: ...
|
|
201
|
+
|
|
202
|
+
|