oci-aivision 2.100.0 → 2.100.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/aiservicevision-waiter.d.ts +8 -0
- package/lib/aiservicevision-waiter.js +12 -0
- package/lib/aiservicevision-waiter.js.map +1 -1
- package/lib/client.d.ts +30 -0
- package/lib/client.js +184 -0
- package/lib/client.js.map +1 -1
- package/lib/model/analyze-video-result.d.ts +71 -0
- package/lib/model/analyze-video-result.js +127 -0
- package/lib/model/analyze-video-result.js.map +1 -0
- package/lib/model/create-video-job-details.d.ts +54 -0
- package/lib/model/create-video-job-details.js +73 -0
- package/lib/model/create-video-job-details.js.map +1 -0
- package/lib/model/index.d.ts +54 -0
- package/lib/model/index.js +57 -2
- package/lib/model/index.js.map +1 -1
- package/lib/model/object-property.d.ts +31 -0
- package/lib/model/object-property.js +57 -0
- package/lib/model/object-property.js.map +1 -0
- package/lib/model/video-face-detection-feature.d.ts +38 -0
- package/lib/model/video-face-detection-feature.js +54 -0
- package/lib/model/video-face-detection-feature.js.map +1 -0
- package/lib/model/video-face-frame.d.ts +39 -0
- package/lib/model/video-face-frame.js +67 -0
- package/lib/model/video-face-frame.js.map +1 -0
- package/lib/model/video-face-segment.d.ts +31 -0
- package/lib/model/video-face-segment.js +67 -0
- package/lib/model/video-face-segment.js.map +1 -0
- package/lib/model/video-face.d.ts +26 -0
- package/lib/model/video-face.js +61 -0
- package/lib/model/video-face.js.map +1 -0
- package/lib/model/video-feature.d.ts +22 -0
- package/lib/model/video-feature.js +84 -0
- package/lib/model/video-feature.js.map +1 -0
- package/lib/model/video-job.d.ts +114 -0
- package/lib/model/video-job.js +97 -0
- package/lib/model/video-job.js.map +1 -0
- package/lib/model/video-label-detection-feature.d.ts +38 -0
- package/lib/model/video-label-detection-feature.js +54 -0
- package/lib/model/video-label-detection-feature.js.map +1 -0
- package/lib/model/video-label-segment.d.ts +27 -0
- package/lib/model/video-label-segment.js +57 -0
- package/lib/model/video-label-segment.js.map +1 -0
- package/lib/model/video-label.d.ts +30 -0
- package/lib/model/video-label.js +61 -0
- package/lib/model/video-label.js.map +1 -0
- package/lib/model/video-metadata.d.ts +37 -0
- package/lib/model/video-metadata.js +29 -0
- package/lib/model/video-metadata.js.map +1 -0
- package/lib/model/video-object-detection-feature.d.ts +38 -0
- package/lib/model/video-object-detection-feature.js +54 -0
- package/lib/model/video-object-detection-feature.js.map +1 -0
- package/lib/model/video-object-frame.d.ts +31 -0
- package/lib/model/video-object-frame.js +57 -0
- package/lib/model/video-object-frame.js.map +1 -0
- package/lib/model/video-object-segment.d.ts +31 -0
- package/lib/model/video-object-segment.js +67 -0
- package/lib/model/video-object-segment.js.map +1 -0
- package/lib/model/video-object-tracking-feature.d.ts +38 -0
- package/lib/model/video-object-tracking-feature.js +54 -0
- package/lib/model/video-object-tracking-feature.js.map +1 -0
- package/lib/model/video-object.d.ts +30 -0
- package/lib/model/video-object.js +61 -0
- package/lib/model/video-object.js.map +1 -0
- package/lib/model/video-segment.d.ts +29 -0
- package/lib/model/video-segment.js +29 -0
- package/lib/model/video-segment.js.map +1 -0
- package/lib/model/video-text-detection-feature.d.ts +30 -0
- package/lib/model/video-text-detection-feature.js +54 -0
- package/lib/model/video-text-detection-feature.js.map +1 -0
- package/lib/model/video-text-frame.d.ts +31 -0
- package/lib/model/video-text-frame.js +57 -0
- package/lib/model/video-text-frame.js.map +1 -0
- package/lib/model/video-text-segment.d.ts +31 -0
- package/lib/model/video-text-segment.js +67 -0
- package/lib/model/video-text-segment.js.map +1 -0
- package/lib/model/video-text.d.ts +30 -0
- package/lib/model/video-text.js +61 -0
- package/lib/model/video-text.js.map +1 -0
- package/lib/model/video-tracked-object-properties.d.ts +25 -0
- package/lib/model/video-tracked-object-properties.js +29 -0
- package/lib/model/video-tracked-object-properties.js.map +1 -0
- package/lib/model/video-tracked-object-segment.d.ts +31 -0
- package/lib/model/video-tracked-object-segment.js +67 -0
- package/lib/model/video-tracked-object-segment.js.map +1 -0
- package/lib/model/video-tracked-object.d.ts +35 -0
- package/lib/model/video-tracked-object.js +67 -0
- package/lib/model/video-tracked-object.js.map +1 -0
- package/lib/model/video-tracking-frame.d.ts +35 -0
- package/lib/model/video-tracking-frame.js +67 -0
- package/lib/model/video-tracking-frame.js.map +1 -0
- package/lib/request/cancel-video-job-request.d.ts +35 -0
- package/lib/request/cancel-video-job-request.js +15 -0
- package/lib/request/cancel-video-job-request.js.map +1 -0
- package/lib/request/create-video-job-request.d.ts +36 -0
- package/lib/request/create-video-job-request.js +15 -0
- package/lib/request/create-video-job-request.js.map +1 -0
- package/lib/request/get-video-job-request.d.ts +26 -0
- package/lib/request/get-video-job-request.js +15 -0
- package/lib/request/get-video-job-request.js.map +1 -0
- package/lib/request/index.d.ts +6 -0
- package/lib/request/index.js.map +1 -1
- package/lib/response/cancel-video-job-response.d.ts +20 -0
- package/lib/response/cancel-video-job-response.js +15 -0
- package/lib/response/cancel-video-job-response.js.map +1 -0
- package/lib/response/create-video-job-response.d.ts +30 -0
- package/lib/response/create-video-job-response.js +15 -0
- package/lib/response/create-video-job-response.js.map +1 -0
- package/lib/response/get-video-job-response.d.ts +30 -0
- package/lib/response/get-video-job-response.js +15 -0
- package/lib/response/get-video-job-response.js.map +1 -0
- package/lib/response/index.d.ts +6 -0
- package/package.json +3 -3
|
@@ -0,0 +1,67 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoTextSegment = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoTextSegment;
|
|
37
|
+
(function (VideoTextSegment) {
|
|
38
|
+
function getJsonObj(obj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
40
|
+
"videoSegment": obj.videoSegment
|
|
41
|
+
? model.VideoSegment.getJsonObj(obj.videoSegment)
|
|
42
|
+
: undefined,
|
|
43
|
+
"frames": obj.frames
|
|
44
|
+
? obj.frames.map(item => {
|
|
45
|
+
return model.VideoTextFrame.getJsonObj(item);
|
|
46
|
+
})
|
|
47
|
+
: undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
VideoTextSegment.getJsonObj = getJsonObj;
|
|
52
|
+
function getDeserializedJsonObj(obj) {
|
|
53
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
54
|
+
"videoSegment": obj.videoSegment
|
|
55
|
+
? model.VideoSegment.getDeserializedJsonObj(obj.videoSegment)
|
|
56
|
+
: undefined,
|
|
57
|
+
"frames": obj.frames
|
|
58
|
+
? obj.frames.map(item => {
|
|
59
|
+
return model.VideoTextFrame.getDeserializedJsonObj(item);
|
|
60
|
+
})
|
|
61
|
+
: undefined
|
|
62
|
+
});
|
|
63
|
+
return jsonObj;
|
|
64
|
+
}
|
|
65
|
+
VideoTextSegment.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
66
|
+
})(VideoTextSegment = exports.VideoTextSegment || (exports.VideoTextSegment = {}));
|
|
67
|
+
//# sourceMappingURL=video-text-segment.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-text-segment.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-text-segment.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAkBlC,IAAiB,gBAAgB,CAqChC;AArCD,WAAiB,gBAAgB;IAC/B,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,cAAc,EAAE,GAAG,CAAC,YAAY;gBAC9B,CAAC,CAAC,KAAK,CAAC,YAAY,CAAC,UAAU,CAAC,GAAG,CAAC,YAAY,CAAC;gBACjD,CAAC,CAAC,SAAS;YAEb,QAAQ,EAAE,GAAG,CAAC,MAAM;gBAClB,CAAC,CAAC,GAAG,CAAC,MAAM,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACpB,OAAO,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC/C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAjBe,2BAAU,aAiBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,cAAc,EAAE,GAAG,CAAC,YAAY;gBAC9B,CAAC,CAAC,KAAK,CAAC,YAAY,CAAC,sBAAsB,CAAC,GAAG,CAAC,YAAY,CAAC;gBAC7D,CAAC,CAAC,SAAS;YAEb,QAAQ,EAAE,GAAG,CAAC,MAAM;gBAClB,CAAC,CAAC,GAAG,CAAC,MAAM,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACpB,OAAO,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC3D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAjBe,uCAAsB,yBAiBrC,CAAA;AACH,CAAC,EArCgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QAqChC"}
|
|
@@ -0,0 +1,30 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
/**
|
|
15
|
+
* Detected text in a video.
|
|
16
|
+
*/
|
|
17
|
+
export interface VideoText {
|
|
18
|
+
/**
|
|
19
|
+
* Detected text.
|
|
20
|
+
*/
|
|
21
|
+
"text": string;
|
|
22
|
+
/**
|
|
23
|
+
* Text segments in a video.
|
|
24
|
+
*/
|
|
25
|
+
"segments": Array<model.VideoTextSegment>;
|
|
26
|
+
}
|
|
27
|
+
export declare namespace VideoText {
|
|
28
|
+
function getJsonObj(obj: VideoText): object;
|
|
29
|
+
function getDeserializedJsonObj(obj: VideoText): object;
|
|
30
|
+
}
|
|
@@ -0,0 +1,61 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoText = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoText;
|
|
37
|
+
(function (VideoText) {
|
|
38
|
+
function getJsonObj(obj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
40
|
+
"segments": obj.segments
|
|
41
|
+
? obj.segments.map(item => {
|
|
42
|
+
return model.VideoTextSegment.getJsonObj(item);
|
|
43
|
+
})
|
|
44
|
+
: undefined
|
|
45
|
+
});
|
|
46
|
+
return jsonObj;
|
|
47
|
+
}
|
|
48
|
+
VideoText.getJsonObj = getJsonObj;
|
|
49
|
+
function getDeserializedJsonObj(obj) {
|
|
50
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
51
|
+
"segments": obj.segments
|
|
52
|
+
? obj.segments.map(item => {
|
|
53
|
+
return model.VideoTextSegment.getDeserializedJsonObj(item);
|
|
54
|
+
})
|
|
55
|
+
: undefined
|
|
56
|
+
});
|
|
57
|
+
return jsonObj;
|
|
58
|
+
}
|
|
59
|
+
VideoText.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
60
|
+
})(VideoText = exports.VideoText || (exports.VideoText = {}));
|
|
61
|
+
//# sourceMappingURL=video-text.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-text.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-text.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAiBlC,IAAiB,SAAS,CA6BzB;AA7BD,WAAiB,SAAS;IACxB,SAAgB,UAAU,CAAC,GAAc;QACvC,MAAM,OAAO,mCACR,GAAG,GACH;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACjD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,oBAAU,aAazB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAc;QACnD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC7D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,gCAAsB,yBAarC,CAAA;AACH,CAAC,EA7BgB,SAAS,GAAT,iBAAS,KAAT,iBAAS,QA6BzB"}
|
|
@@ -0,0 +1,25 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
/**
|
|
14
|
+
* Properties of a tracked object in a video.
|
|
15
|
+
*/
|
|
16
|
+
export interface VideoTrackedObjectProperties {
|
|
17
|
+
/**
|
|
18
|
+
* The axle count value of a tracked vehicle. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
19
|
+
*/
|
|
20
|
+
"axleCount"?: number;
|
|
21
|
+
}
|
|
22
|
+
export declare namespace VideoTrackedObjectProperties {
|
|
23
|
+
function getJsonObj(obj: VideoTrackedObjectProperties): object;
|
|
24
|
+
function getDeserializedJsonObj(obj: VideoTrackedObjectProperties): object;
|
|
25
|
+
}
|
|
@@ -0,0 +1,29 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
15
|
+
exports.VideoTrackedObjectProperties = void 0;
|
|
16
|
+
var VideoTrackedObjectProperties;
|
|
17
|
+
(function (VideoTrackedObjectProperties) {
|
|
18
|
+
function getJsonObj(obj) {
|
|
19
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
20
|
+
return jsonObj;
|
|
21
|
+
}
|
|
22
|
+
VideoTrackedObjectProperties.getJsonObj = getJsonObj;
|
|
23
|
+
function getDeserializedJsonObj(obj) {
|
|
24
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
25
|
+
return jsonObj;
|
|
26
|
+
}
|
|
27
|
+
VideoTrackedObjectProperties.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
28
|
+
})(VideoTrackedObjectProperties = exports.VideoTrackedObjectProperties || (exports.VideoTrackedObjectProperties = {}));
|
|
29
|
+
//# sourceMappingURL=video-tracked-object-properties.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-tracked-object-properties.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-tracked-object-properties.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;AAeH,IAAiB,4BAA4B,CAW5C;AAXD,WAAiB,4BAA4B;IAC3C,SAAgB,UAAU,CAAC,GAAiC;QAC1D,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,uCAAU,aAIzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAiC;QACtE,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,OAAO,OAAO,CAAC;IACjB,CAAC;IAJe,mDAAsB,yBAIrC,CAAA;AACH,CAAC,EAXgB,4BAA4B,GAA5B,oCAA4B,KAA5B,oCAA4B,QAW5C"}
|
|
@@ -0,0 +1,31 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
/**
|
|
15
|
+
* A segment of a tracked object in a video.
|
|
16
|
+
*/
|
|
17
|
+
export interface VideoTrackedObjectSegment {
|
|
18
|
+
"videoSegment": model.VideoSegment;
|
|
19
|
+
/**
|
|
20
|
+
* The confidence score, between 0 and 1. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
21
|
+
*/
|
|
22
|
+
"confidence": number;
|
|
23
|
+
/**
|
|
24
|
+
* Frames within the segment.
|
|
25
|
+
*/
|
|
26
|
+
"frames": Array<model.VideoTrackingFrame>;
|
|
27
|
+
}
|
|
28
|
+
export declare namespace VideoTrackedObjectSegment {
|
|
29
|
+
function getJsonObj(obj: VideoTrackedObjectSegment): object;
|
|
30
|
+
function getDeserializedJsonObj(obj: VideoTrackedObjectSegment): object;
|
|
31
|
+
}
|
|
@@ -0,0 +1,67 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoTrackedObjectSegment = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoTrackedObjectSegment;
|
|
37
|
+
(function (VideoTrackedObjectSegment) {
|
|
38
|
+
function getJsonObj(obj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
40
|
+
"videoSegment": obj.videoSegment
|
|
41
|
+
? model.VideoSegment.getJsonObj(obj.videoSegment)
|
|
42
|
+
: undefined,
|
|
43
|
+
"frames": obj.frames
|
|
44
|
+
? obj.frames.map(item => {
|
|
45
|
+
return model.VideoTrackingFrame.getJsonObj(item);
|
|
46
|
+
})
|
|
47
|
+
: undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
VideoTrackedObjectSegment.getJsonObj = getJsonObj;
|
|
52
|
+
function getDeserializedJsonObj(obj) {
|
|
53
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
54
|
+
"videoSegment": obj.videoSegment
|
|
55
|
+
? model.VideoSegment.getDeserializedJsonObj(obj.videoSegment)
|
|
56
|
+
: undefined,
|
|
57
|
+
"frames": obj.frames
|
|
58
|
+
? obj.frames.map(item => {
|
|
59
|
+
return model.VideoTrackingFrame.getDeserializedJsonObj(item);
|
|
60
|
+
})
|
|
61
|
+
: undefined
|
|
62
|
+
});
|
|
63
|
+
return jsonObj;
|
|
64
|
+
}
|
|
65
|
+
VideoTrackedObjectSegment.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
66
|
+
})(VideoTrackedObjectSegment = exports.VideoTrackedObjectSegment || (exports.VideoTrackedObjectSegment = {}));
|
|
67
|
+
//# sourceMappingURL=video-tracked-object-segment.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-tracked-object-segment.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-tracked-object-segment.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAkBlC,IAAiB,yBAAyB,CAqCzC;AArCD,WAAiB,yBAAyB;IACxC,SAAgB,UAAU,CAAC,GAA8B;QACvD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,cAAc,EAAE,GAAG,CAAC,YAAY;gBAC9B,CAAC,CAAC,KAAK,CAAC,YAAY,CAAC,UAAU,CAAC,GAAG,CAAC,YAAY,CAAC;gBACjD,CAAC,CAAC,SAAS;YAEb,QAAQ,EAAE,GAAG,CAAC,MAAM;gBAClB,CAAC,CAAC,GAAG,CAAC,MAAM,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACpB,OAAO,KAAK,CAAC,kBAAkB,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACnD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAjBe,oCAAU,aAiBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAA8B;QACnE,MAAM,OAAO,mCACR,GAAG,GACH;YACD,cAAc,EAAE,GAAG,CAAC,YAAY;gBAC9B,CAAC,CAAC,KAAK,CAAC,YAAY,CAAC,sBAAsB,CAAC,GAAG,CAAC,YAAY,CAAC;gBAC7D,CAAC,CAAC,SAAS;YAEb,QAAQ,EAAE,GAAG,CAAC,MAAM;gBAClB,CAAC,CAAC,GAAG,CAAC,MAAM,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACpB,OAAO,KAAK,CAAC,kBAAkB,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC/D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAjBe,gDAAsB,yBAiBrC,CAAA;AACH,CAAC,EArCgB,yBAAyB,GAAzB,iCAAyB,KAAzB,iCAAyB,QAqCzC"}
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
/**
|
|
15
|
+
* Tracked object in a video.
|
|
16
|
+
*/
|
|
17
|
+
export interface VideoTrackedObject {
|
|
18
|
+
/**
|
|
19
|
+
* Name of the object category label.
|
|
20
|
+
*/
|
|
21
|
+
"name": string;
|
|
22
|
+
/**
|
|
23
|
+
* Unique identifier for the object. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
24
|
+
*/
|
|
25
|
+
"objectId": number;
|
|
26
|
+
"properties"?: model.VideoTrackedObjectProperties;
|
|
27
|
+
/**
|
|
28
|
+
* Segments for the tracked object.
|
|
29
|
+
*/
|
|
30
|
+
"segments": Array<model.VideoTrackedObjectSegment>;
|
|
31
|
+
}
|
|
32
|
+
export declare namespace VideoTrackedObject {
|
|
33
|
+
function getJsonObj(obj: VideoTrackedObject): object;
|
|
34
|
+
function getDeserializedJsonObj(obj: VideoTrackedObject): object;
|
|
35
|
+
}
|
|
@@ -0,0 +1,67 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoTrackedObject = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoTrackedObject;
|
|
37
|
+
(function (VideoTrackedObject) {
|
|
38
|
+
function getJsonObj(obj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
40
|
+
"properties": obj.properties
|
|
41
|
+
? model.VideoTrackedObjectProperties.getJsonObj(obj.properties)
|
|
42
|
+
: undefined,
|
|
43
|
+
"segments": obj.segments
|
|
44
|
+
? obj.segments.map(item => {
|
|
45
|
+
return model.VideoTrackedObjectSegment.getJsonObj(item);
|
|
46
|
+
})
|
|
47
|
+
: undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
VideoTrackedObject.getJsonObj = getJsonObj;
|
|
52
|
+
function getDeserializedJsonObj(obj) {
|
|
53
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
54
|
+
"properties": obj.properties
|
|
55
|
+
? model.VideoTrackedObjectProperties.getDeserializedJsonObj(obj.properties)
|
|
56
|
+
: undefined,
|
|
57
|
+
"segments": obj.segments
|
|
58
|
+
? obj.segments.map(item => {
|
|
59
|
+
return model.VideoTrackedObjectSegment.getDeserializedJsonObj(item);
|
|
60
|
+
})
|
|
61
|
+
: undefined
|
|
62
|
+
});
|
|
63
|
+
return jsonObj;
|
|
64
|
+
}
|
|
65
|
+
VideoTrackedObject.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
66
|
+
})(VideoTrackedObject = exports.VideoTrackedObject || (exports.VideoTrackedObject = {}));
|
|
67
|
+
//# sourceMappingURL=video-tracked-object.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-tracked-object.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-tracked-object.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAsBlC,IAAiB,kBAAkB,CAmClC;AAnCD,WAAiB,kBAAkB;IACjC,SAAgB,UAAU,CAAC,GAAuB;QAChD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,KAAK,CAAC,4BAA4B,CAAC,UAAU,CAAC,GAAG,CAAC,UAAU,CAAC;gBAC/D,CAAC,CAAC,SAAS;YACb,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,yBAAyB,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC1D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhBe,6BAAU,aAgBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAuB;QAC5D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,KAAK,CAAC,4BAA4B,CAAC,sBAAsB,CAAC,GAAG,CAAC,UAAU,CAAC;gBAC3E,CAAC,CAAC,SAAS;YACb,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,yBAAyB,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACtE,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhBe,yCAAsB,yBAgBrC,CAAA;AACH,CAAC,EAnCgB,kBAAkB,GAAlB,0BAAkB,KAAlB,0BAAkB,QAmClC"}
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
/**
|
|
15
|
+
* A frame capturing a tracked object.
|
|
16
|
+
*/
|
|
17
|
+
export interface VideoTrackingFrame {
|
|
18
|
+
/**
|
|
19
|
+
* Time offset(Milliseconds) of the frame. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
20
|
+
*/
|
|
21
|
+
"timeOffsetMs": number;
|
|
22
|
+
/**
|
|
23
|
+
* The confidence score, between 0 and 1. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
24
|
+
*/
|
|
25
|
+
"confidence": number;
|
|
26
|
+
"boundingPolygon": model.BoundingPolygon;
|
|
27
|
+
/**
|
|
28
|
+
* Properties associated with the tracked object in the frame.
|
|
29
|
+
*/
|
|
30
|
+
"properties"?: Array<model.ObjectProperty>;
|
|
31
|
+
}
|
|
32
|
+
export declare namespace VideoTrackingFrame {
|
|
33
|
+
function getJsonObj(obj: VideoTrackingFrame): object;
|
|
34
|
+
function getDeserializedJsonObj(obj: VideoTrackingFrame): object;
|
|
35
|
+
}
|
|
@@ -0,0 +1,67 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoTrackingFrame = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoTrackingFrame;
|
|
37
|
+
(function (VideoTrackingFrame) {
|
|
38
|
+
function getJsonObj(obj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
40
|
+
"boundingPolygon": obj.boundingPolygon
|
|
41
|
+
? model.BoundingPolygon.getJsonObj(obj.boundingPolygon)
|
|
42
|
+
: undefined,
|
|
43
|
+
"properties": obj.properties
|
|
44
|
+
? obj.properties.map(item => {
|
|
45
|
+
return model.ObjectProperty.getJsonObj(item);
|
|
46
|
+
})
|
|
47
|
+
: undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
VideoTrackingFrame.getJsonObj = getJsonObj;
|
|
52
|
+
function getDeserializedJsonObj(obj) {
|
|
53
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
54
|
+
"boundingPolygon": obj.boundingPolygon
|
|
55
|
+
? model.BoundingPolygon.getDeserializedJsonObj(obj.boundingPolygon)
|
|
56
|
+
: undefined,
|
|
57
|
+
"properties": obj.properties
|
|
58
|
+
? obj.properties.map(item => {
|
|
59
|
+
return model.ObjectProperty.getDeserializedJsonObj(item);
|
|
60
|
+
})
|
|
61
|
+
: undefined
|
|
62
|
+
});
|
|
63
|
+
return jsonObj;
|
|
64
|
+
}
|
|
65
|
+
VideoTrackingFrame.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
66
|
+
})(VideoTrackingFrame = exports.VideoTrackingFrame || (exports.VideoTrackingFrame = {}));
|
|
67
|
+
//# sourceMappingURL=video-tracking-frame.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-tracking-frame.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-tracking-frame.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAsBlC,IAAiB,kBAAkB,CAmClC;AAnCD,WAAiB,kBAAkB;IACjC,SAAgB,UAAU,CAAC,GAAuB;QAChD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAC,eAAe,CAAC;gBACvD,CAAC,CAAC,SAAS;YACb,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,GAAG,CAAC,UAAU,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACxB,OAAO,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC/C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhBe,6BAAU,aAgBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAuB;QAC5D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAC,eAAe,CAAC;gBACnE,CAAC,CAAC,SAAS;YACb,YAAY,EAAE,GAAG,CAAC,UAAU;gBAC1B,CAAC,CAAC,GAAG,CAAC,UAAU,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACxB,OAAO,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC3D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhBe,yCAAsB,yBAgBrC,CAAA;AACH,CAAC,EAnCgB,kBAAkB,GAAlB,0BAAkB,KAAlB,0BAAkB,QAmClC"}
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
/**
|
|
2
|
+
*
|
|
3
|
+
*
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import common = require("oci-common");
|
|
14
|
+
/**
|
|
15
|
+
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/latest/aivision/CancelVideoJob.ts.html |here} to see how to use CancelVideoJobRequest.
|
|
16
|
+
*/
|
|
17
|
+
export interface CancelVideoJobRequest extends common.BaseRequest {
|
|
18
|
+
/**
|
|
19
|
+
* Video job id.
|
|
20
|
+
*/
|
|
21
|
+
"videoJobId": string;
|
|
22
|
+
/**
|
|
23
|
+
* For optimistic concurrency control. In the PUT or DELETE call
|
|
24
|
+
* for a resource, set the {@code if-match} parameter to the value of the
|
|
25
|
+
* etag from a previous GET or POST response for that resource.
|
|
26
|
+
* The resource will be updated or deleted only if the etag you
|
|
27
|
+
* provide matches the resource's current etag value.
|
|
28
|
+
*
|
|
29
|
+
*/
|
|
30
|
+
"ifMatch"?: string;
|
|
31
|
+
/**
|
|
32
|
+
* The client request ID for tracing.
|
|
33
|
+
*/
|
|
34
|
+
"opcRequestId"?: string;
|
|
35
|
+
}
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
*
|
|
4
|
+
*
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
15
|
+
//# sourceMappingURL=cancel-video-job-request.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cancel-video-job-request.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/request/cancel-video-job-request.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG"}
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
/**
|
|
2
|
+
*
|
|
3
|
+
*
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
import common = require("oci-common");
|
|
15
|
+
/**
|
|
16
|
+
* @example Click {@link https://docs.cloud.oracle.com/en-us/iaas/tools/typescript-sdk-examples/latest/aivision/CreateVideoJob.ts.html |here} to see how to use CreateVideoJobRequest.
|
|
17
|
+
*/
|
|
18
|
+
export interface CreateVideoJobRequest extends common.BaseRequest {
|
|
19
|
+
/**
|
|
20
|
+
* Details about the video analysis.
|
|
21
|
+
*/
|
|
22
|
+
"createVideoJobDetails": model.CreateVideoJobDetails;
|
|
23
|
+
/**
|
|
24
|
+
* A token that uniquely identifies a request so it can be retried in case of a timeout or
|
|
25
|
+
* server error without the risk of executing that same action again. Retry tokens expire after 24
|
|
26
|
+
* hours, but can be invalidated before then due to conflicting operations. For example, if a resource
|
|
27
|
+
* has been deleted and purged from the system, then a retry of the original creation request
|
|
28
|
+
* might be rejected.
|
|
29
|
+
*
|
|
30
|
+
*/
|
|
31
|
+
"opcRetryToken"?: string;
|
|
32
|
+
/**
|
|
33
|
+
* The client request ID for tracing.
|
|
34
|
+
*/
|
|
35
|
+
"opcRequestId"?: string;
|
|
36
|
+
}
|