oci-aivision 2.99.0 → 2.100.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/aiservicevision-waiter.d.ts +8 -0
- package/lib/aiservicevision-waiter.js +12 -0
- package/lib/aiservicevision-waiter.js.map +1 -1
- package/lib/client.d.ts +35 -0
- package/lib/client.js +195 -0
- package/lib/client.js.map +1 -1
- package/lib/model/analyze-video-result.d.ts +71 -0
- package/lib/model/analyze-video-result.js +127 -0
- package/lib/model/analyze-video-result.js.map +1 -0
- package/lib/model/create-video-job-details.d.ts +54 -0
- package/lib/model/create-video-job-details.js +73 -0
- package/lib/model/create-video-job-details.js.map +1 -0
- package/lib/model/index.d.ts +54 -0
- package/lib/model/index.js +57 -2
- package/lib/model/index.js.map +1 -1
- package/lib/model/object-property.d.ts +31 -0
- package/lib/model/object-property.js +57 -0
- package/lib/model/object-property.js.map +1 -0
- package/lib/model/video-face-detection-feature.d.ts +38 -0
- package/lib/model/video-face-detection-feature.js +54 -0
- package/lib/model/video-face-detection-feature.js.map +1 -0
- package/lib/model/video-face-frame.d.ts +39 -0
- package/lib/model/video-face-frame.js +67 -0
- package/lib/model/video-face-frame.js.map +1 -0
- package/lib/model/video-face-segment.d.ts +31 -0
- package/lib/model/video-face-segment.js +67 -0
- package/lib/model/video-face-segment.js.map +1 -0
- package/lib/model/video-face.d.ts +26 -0
- package/lib/model/video-face.js +61 -0
- package/lib/model/video-face.js.map +1 -0
- package/lib/model/video-feature.d.ts +22 -0
- package/lib/model/video-feature.js +84 -0
- package/lib/model/video-feature.js.map +1 -0
- package/lib/model/video-job.d.ts +114 -0
- package/lib/model/video-job.js +97 -0
- package/lib/model/video-job.js.map +1 -0
- package/lib/model/video-label-detection-feature.d.ts +38 -0
- package/lib/model/video-label-detection-feature.js +54 -0
- package/lib/model/video-label-detection-feature.js.map +1 -0
- package/lib/model/video-label-segment.d.ts +27 -0
- package/lib/model/video-label-segment.js +57 -0
- package/lib/model/video-label-segment.js.map +1 -0
- package/lib/model/video-label.d.ts +30 -0
- package/lib/model/video-label.js +61 -0
- package/lib/model/video-label.js.map +1 -0
- package/lib/model/video-metadata.d.ts +37 -0
- package/lib/model/video-metadata.js +29 -0
- package/lib/model/video-metadata.js.map +1 -0
- package/lib/model/video-object-detection-feature.d.ts +38 -0
- package/lib/model/video-object-detection-feature.js +54 -0
- package/lib/model/video-object-detection-feature.js.map +1 -0
- package/lib/model/video-object-frame.d.ts +31 -0
- package/lib/model/video-object-frame.js +57 -0
- package/lib/model/video-object-frame.js.map +1 -0
- package/lib/model/video-object-segment.d.ts +31 -0
- package/lib/model/video-object-segment.js +67 -0
- package/lib/model/video-object-segment.js.map +1 -0
- package/lib/model/video-object-tracking-feature.d.ts +38 -0
- package/lib/model/video-object-tracking-feature.js +54 -0
- package/lib/model/video-object-tracking-feature.js.map +1 -0
- package/lib/model/video-object.d.ts +30 -0
- package/lib/model/video-object.js +61 -0
- package/lib/model/video-object.js.map +1 -0
- package/lib/model/video-segment.d.ts +29 -0
- package/lib/model/video-segment.js +29 -0
- package/lib/model/video-segment.js.map +1 -0
- package/lib/model/video-text-detection-feature.d.ts +30 -0
- package/lib/model/video-text-detection-feature.js +54 -0
- package/lib/model/video-text-detection-feature.js.map +1 -0
- package/lib/model/video-text-frame.d.ts +31 -0
- package/lib/model/video-text-frame.js +57 -0
- package/lib/model/video-text-frame.js.map +1 -0
- package/lib/model/video-text-segment.d.ts +31 -0
- package/lib/model/video-text-segment.js +67 -0
- package/lib/model/video-text-segment.js.map +1 -0
- package/lib/model/video-text.d.ts +30 -0
- package/lib/model/video-text.js +61 -0
- package/lib/model/video-text.js.map +1 -0
- package/lib/model/video-tracked-object-properties.d.ts +25 -0
- package/lib/model/video-tracked-object-properties.js +29 -0
- package/lib/model/video-tracked-object-properties.js.map +1 -0
- package/lib/model/video-tracked-object-segment.d.ts +31 -0
- package/lib/model/video-tracked-object-segment.js +67 -0
- package/lib/model/video-tracked-object-segment.js.map +1 -0
- package/lib/model/video-tracked-object.d.ts +35 -0
- package/lib/model/video-tracked-object.js +67 -0
- package/lib/model/video-tracked-object.js.map +1 -0
- package/lib/model/video-tracking-frame.d.ts +35 -0
- package/lib/model/video-tracking-frame.js +67 -0
- package/lib/model/video-tracking-frame.js.map +1 -0
- package/lib/request/cancel-video-job-request.d.ts +35 -0
- package/lib/request/cancel-video-job-request.js +15 -0
- package/lib/request/cancel-video-job-request.js.map +1 -0
- package/lib/request/create-video-job-request.d.ts +36 -0
- package/lib/request/create-video-job-request.js +15 -0
- package/lib/request/create-video-job-request.js.map +1 -0
- package/lib/request/get-video-job-request.d.ts +26 -0
- package/lib/request/get-video-job-request.js +15 -0
- package/lib/request/get-video-job-request.js.map +1 -0
- package/lib/request/index.d.ts +6 -0
- package/lib/request/index.js.map +1 -1
- package/lib/response/cancel-video-job-response.d.ts +20 -0
- package/lib/response/cancel-video-job-response.js +15 -0
- package/lib/response/cancel-video-job-response.js.map +1 -0
- package/lib/response/create-video-job-response.d.ts +30 -0
- package/lib/response/create-video-job-response.js +15 -0
- package/lib/response/create-video-job-response.js.map +1 -0
- package/lib/response/get-video-job-response.d.ts +30 -0
- package/lib/response/get-video-job-response.js +15 -0
- package/lib/response/get-video-job-response.js.map +1 -0
- package/lib/response/index.d.ts +6 -0
- package/package.json +3 -3
|
@@ -0,0 +1,54 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoFaceDetectionFeature = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoFaceDetectionFeature;
|
|
37
|
+
(function (VideoFaceDetectionFeature) {
|
|
38
|
+
function getJsonObj(obj, isParentJsonObj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
40
|
+
? obj
|
|
41
|
+
: model.VideoFeature.getJsonObj(obj))), {});
|
|
42
|
+
return jsonObj;
|
|
43
|
+
}
|
|
44
|
+
VideoFaceDetectionFeature.getJsonObj = getJsonObj;
|
|
45
|
+
VideoFaceDetectionFeature.featureType = "FACE_DETECTION";
|
|
46
|
+
function getDeserializedJsonObj(obj, isParentJsonObj) {
|
|
47
|
+
const jsonObj = Object.assign(Object.assign({}, (isParentJsonObj
|
|
48
|
+
? obj
|
|
49
|
+
: model.VideoFeature.getDeserializedJsonObj(obj))), {});
|
|
50
|
+
return jsonObj;
|
|
51
|
+
}
|
|
52
|
+
VideoFaceDetectionFeature.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
53
|
+
})(VideoFaceDetectionFeature = exports.VideoFaceDetectionFeature || (exports.VideoFaceDetectionFeature = {}));
|
|
54
|
+
//# sourceMappingURL=video-face-detection-feature.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-face-detection-feature.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-face-detection-feature.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAyBlC,IAAiB,yBAAyB,CAyBzC;AAzBD,WAAiB,yBAAyB;IACxC,SAAgB,UAAU,CAAC,GAA8B,EAAE,eAAyB;QAClF,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,YAAY,CAAC,UAAU,CAAC,GAAG,CAA+B,CAAC,GACnE,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IATe,oCAAU,aASzB,CAAA;IACY,qCAAW,GAAG,gBAAgB,CAAC;IAC5C,SAAgB,sBAAsB,CACpC,GAA8B,EAC9B,eAAyB;QAEzB,MAAM,OAAO,mCACR,CAAC,eAAe;YACjB,CAAC,CAAC,GAAG;YACL,CAAC,CAAE,KAAK,CAAC,YAAY,CAAC,sBAAsB,CAAC,GAAG,CAA+B,CAAC,GAC/E,EAAE,CACN,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAZe,gDAAsB,yBAYrC,CAAA;AACH,CAAC,EAzBgB,yBAAyB,GAAzB,iCAAyB,KAAzB,iCAAyB,QAyBzC"}
|
|
@@ -0,0 +1,39 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
/**
|
|
15
|
+
* A face frame.
|
|
16
|
+
*/
|
|
17
|
+
export interface VideoFaceFrame {
|
|
18
|
+
/**
|
|
19
|
+
* Time offset(Milliseconds) in the video. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
20
|
+
*/
|
|
21
|
+
"timeOffsetMs": number;
|
|
22
|
+
/**
|
|
23
|
+
* The confidence score, between 0 and 1. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
24
|
+
*/
|
|
25
|
+
"confidence": number;
|
|
26
|
+
/**
|
|
27
|
+
* The quality score of the face detected, between 0 and 1. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
28
|
+
*/
|
|
29
|
+
"qualityScore": number;
|
|
30
|
+
"boundingPolygon": model.BoundingPolygon;
|
|
31
|
+
/**
|
|
32
|
+
* Face landmarks.
|
|
33
|
+
*/
|
|
34
|
+
"landmarks"?: Array<model.Landmark>;
|
|
35
|
+
}
|
|
36
|
+
export declare namespace VideoFaceFrame {
|
|
37
|
+
function getJsonObj(obj: VideoFaceFrame): object;
|
|
38
|
+
function getDeserializedJsonObj(obj: VideoFaceFrame): object;
|
|
39
|
+
}
|
|
@@ -0,0 +1,67 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoFaceFrame = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoFaceFrame;
|
|
37
|
+
(function (VideoFaceFrame) {
|
|
38
|
+
function getJsonObj(obj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
40
|
+
"boundingPolygon": obj.boundingPolygon
|
|
41
|
+
? model.BoundingPolygon.getJsonObj(obj.boundingPolygon)
|
|
42
|
+
: undefined,
|
|
43
|
+
"landmarks": obj.landmarks
|
|
44
|
+
? obj.landmarks.map(item => {
|
|
45
|
+
return model.Landmark.getJsonObj(item);
|
|
46
|
+
})
|
|
47
|
+
: undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
VideoFaceFrame.getJsonObj = getJsonObj;
|
|
52
|
+
function getDeserializedJsonObj(obj) {
|
|
53
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
54
|
+
"boundingPolygon": obj.boundingPolygon
|
|
55
|
+
? model.BoundingPolygon.getDeserializedJsonObj(obj.boundingPolygon)
|
|
56
|
+
: undefined,
|
|
57
|
+
"landmarks": obj.landmarks
|
|
58
|
+
? obj.landmarks.map(item => {
|
|
59
|
+
return model.Landmark.getDeserializedJsonObj(item);
|
|
60
|
+
})
|
|
61
|
+
: undefined
|
|
62
|
+
});
|
|
63
|
+
return jsonObj;
|
|
64
|
+
}
|
|
65
|
+
VideoFaceFrame.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
66
|
+
})(VideoFaceFrame = exports.VideoFaceFrame || (exports.VideoFaceFrame = {}));
|
|
67
|
+
//# sourceMappingURL=video-face-frame.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-face-frame.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-face-frame.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AA0BlC,IAAiB,cAAc,CAmC9B;AAnCD,WAAiB,cAAc;IAC7B,SAAgB,UAAU,CAAC,GAAmB;QAC5C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,KAAK,CAAC,eAAe,CAAC,UAAU,CAAC,GAAG,CAAC,eAAe,CAAC;gBACvD,CAAC,CAAC,SAAS;YACb,WAAW,EAAE,GAAG,CAAC,SAAS;gBACxB,CAAC,CAAC,GAAG,CAAC,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACvB,OAAO,KAAK,CAAC,QAAQ,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACzC,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhBe,yBAAU,aAgBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAmB;QACxD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,iBAAiB,EAAE,GAAG,CAAC,eAAe;gBACpC,CAAC,CAAC,KAAK,CAAC,eAAe,CAAC,sBAAsB,CAAC,GAAG,CAAC,eAAe,CAAC;gBACnE,CAAC,CAAC,SAAS;YACb,WAAW,EAAE,GAAG,CAAC,SAAS;gBACxB,CAAC,CAAC,GAAG,CAAC,SAAS,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACvB,OAAO,KAAK,CAAC,QAAQ,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBACrD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAhBe,qCAAsB,yBAgBrC,CAAA;AACH,CAAC,EAnCgB,cAAc,GAAd,sBAAc,KAAd,sBAAc,QAmC9B"}
|
|
@@ -0,0 +1,31 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
/**
|
|
15
|
+
* A face segment in a video.
|
|
16
|
+
*/
|
|
17
|
+
export interface VideoFaceSegment {
|
|
18
|
+
"videoSegment": model.VideoSegment;
|
|
19
|
+
/**
|
|
20
|
+
* The confidence score, between 0 and 1. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
21
|
+
*/
|
|
22
|
+
"confidence": number;
|
|
23
|
+
/**
|
|
24
|
+
* Face frames in a segment.
|
|
25
|
+
*/
|
|
26
|
+
"frames": Array<model.VideoFaceFrame>;
|
|
27
|
+
}
|
|
28
|
+
export declare namespace VideoFaceSegment {
|
|
29
|
+
function getJsonObj(obj: VideoFaceSegment): object;
|
|
30
|
+
function getDeserializedJsonObj(obj: VideoFaceSegment): object;
|
|
31
|
+
}
|
|
@@ -0,0 +1,67 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoFaceSegment = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoFaceSegment;
|
|
37
|
+
(function (VideoFaceSegment) {
|
|
38
|
+
function getJsonObj(obj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
40
|
+
"videoSegment": obj.videoSegment
|
|
41
|
+
? model.VideoSegment.getJsonObj(obj.videoSegment)
|
|
42
|
+
: undefined,
|
|
43
|
+
"frames": obj.frames
|
|
44
|
+
? obj.frames.map(item => {
|
|
45
|
+
return model.VideoFaceFrame.getJsonObj(item);
|
|
46
|
+
})
|
|
47
|
+
: undefined
|
|
48
|
+
});
|
|
49
|
+
return jsonObj;
|
|
50
|
+
}
|
|
51
|
+
VideoFaceSegment.getJsonObj = getJsonObj;
|
|
52
|
+
function getDeserializedJsonObj(obj) {
|
|
53
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
54
|
+
"videoSegment": obj.videoSegment
|
|
55
|
+
? model.VideoSegment.getDeserializedJsonObj(obj.videoSegment)
|
|
56
|
+
: undefined,
|
|
57
|
+
"frames": obj.frames
|
|
58
|
+
? obj.frames.map(item => {
|
|
59
|
+
return model.VideoFaceFrame.getDeserializedJsonObj(item);
|
|
60
|
+
})
|
|
61
|
+
: undefined
|
|
62
|
+
});
|
|
63
|
+
return jsonObj;
|
|
64
|
+
}
|
|
65
|
+
VideoFaceSegment.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
66
|
+
})(VideoFaceSegment = exports.VideoFaceSegment || (exports.VideoFaceSegment = {}));
|
|
67
|
+
//# sourceMappingURL=video-face-segment.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-face-segment.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-face-segment.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAkBlC,IAAiB,gBAAgB,CAqChC;AArCD,WAAiB,gBAAgB;IAC/B,SAAgB,UAAU,CAAC,GAAqB;QAC9C,MAAM,OAAO,mCACR,GAAG,GACH;YACD,cAAc,EAAE,GAAG,CAAC,YAAY;gBAC9B,CAAC,CAAC,KAAK,CAAC,YAAY,CAAC,UAAU,CAAC,GAAG,CAAC,YAAY,CAAC;gBACjD,CAAC,CAAC,SAAS;YAEb,QAAQ,EAAE,GAAG,CAAC,MAAM;gBAClB,CAAC,CAAC,GAAG,CAAC,MAAM,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACpB,OAAO,KAAK,CAAC,cAAc,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBAC/C,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAjBe,2BAAU,aAiBzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAqB;QAC1D,MAAM,OAAO,mCACR,GAAG,GACH;YACD,cAAc,EAAE,GAAG,CAAC,YAAY;gBAC9B,CAAC,CAAC,KAAK,CAAC,YAAY,CAAC,sBAAsB,CAAC,GAAG,CAAC,YAAY,CAAC;gBAC7D,CAAC,CAAC,SAAS;YAEb,QAAQ,EAAE,GAAG,CAAC,MAAM;gBAClB,CAAC,CAAC,GAAG,CAAC,MAAM,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACpB,OAAO,KAAK,CAAC,cAAc,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC3D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAjBe,uCAAsB,yBAiBrC,CAAA;AACH,CAAC,EArCgB,gBAAgB,GAAhB,wBAAgB,KAAhB,wBAAgB,QAqChC"}
|
|
@@ -0,0 +1,26 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
/**
|
|
15
|
+
* Detected face in a video.
|
|
16
|
+
*/
|
|
17
|
+
export interface VideoFace {
|
|
18
|
+
/**
|
|
19
|
+
* Face segments in a video.
|
|
20
|
+
*/
|
|
21
|
+
"segments": Array<model.VideoFaceSegment>;
|
|
22
|
+
}
|
|
23
|
+
export declare namespace VideoFace {
|
|
24
|
+
function getJsonObj(obj: VideoFace): object;
|
|
25
|
+
function getDeserializedJsonObj(obj: VideoFace): object;
|
|
26
|
+
}
|
|
@@ -0,0 +1,61 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoFace = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
var VideoFace;
|
|
37
|
+
(function (VideoFace) {
|
|
38
|
+
function getJsonObj(obj) {
|
|
39
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
40
|
+
"segments": obj.segments
|
|
41
|
+
? obj.segments.map(item => {
|
|
42
|
+
return model.VideoFaceSegment.getJsonObj(item);
|
|
43
|
+
})
|
|
44
|
+
: undefined
|
|
45
|
+
});
|
|
46
|
+
return jsonObj;
|
|
47
|
+
}
|
|
48
|
+
VideoFace.getJsonObj = getJsonObj;
|
|
49
|
+
function getDeserializedJsonObj(obj) {
|
|
50
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {
|
|
51
|
+
"segments": obj.segments
|
|
52
|
+
? obj.segments.map(item => {
|
|
53
|
+
return model.VideoFaceSegment.getDeserializedJsonObj(item);
|
|
54
|
+
})
|
|
55
|
+
: undefined
|
|
56
|
+
});
|
|
57
|
+
return jsonObj;
|
|
58
|
+
}
|
|
59
|
+
VideoFace.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
60
|
+
})(VideoFace = exports.VideoFace || (exports.VideoFace = {}));
|
|
61
|
+
//# sourceMappingURL=video-face.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-face.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-face.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAalC,IAAiB,SAAS,CA6BzB;AA7BD,WAAiB,SAAS;IACxB,SAAgB,UAAU,CAAC,GAAc;QACvC,MAAM,OAAO,mCACR,GAAG,GACH;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,gBAAgB,CAAC,UAAU,CAAC,IAAI,CAAC,CAAC;gBACjD,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,oBAAU,aAazB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAc;QACnD,MAAM,OAAO,mCACR,GAAG,GACH;YACD,UAAU,EAAE,GAAG,CAAC,QAAQ;gBACtB,CAAC,CAAC,GAAG,CAAC,QAAQ,CAAC,GAAG,CAAC,IAAI,CAAC,EAAE;oBACtB,OAAO,KAAK,CAAC,gBAAgB,CAAC,sBAAsB,CAAC,IAAI,CAAC,CAAC;gBAC7D,CAAC,CAAC;gBACJ,CAAC,CAAC,SAAS;SACd,CACF,CAAC;QAEF,OAAO,OAAO,CAAC;IACjB,CAAC;IAbe,gCAAsB,yBAarC,CAAA;AACH,CAAC,EA7BgB,SAAS,GAAT,iBAAS,KAAT,iBAAS,QA6BzB"}
|
|
@@ -0,0 +1,22 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
/**
|
|
14
|
+
* Details about a video feature request.
|
|
15
|
+
*/
|
|
16
|
+
export interface VideoFeature {
|
|
17
|
+
"featureType": string;
|
|
18
|
+
}
|
|
19
|
+
export declare namespace VideoFeature {
|
|
20
|
+
function getJsonObj(obj: VideoFeature): object;
|
|
21
|
+
function getDeserializedJsonObj(obj: VideoFeature): object;
|
|
22
|
+
}
|
|
@@ -0,0 +1,84 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/**
|
|
3
|
+
* Vision API
|
|
4
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
5
|
+
* OpenAPI spec version: 20220125
|
|
6
|
+
*
|
|
7
|
+
*
|
|
8
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
9
|
+
* Do not edit the class manually.
|
|
10
|
+
*
|
|
11
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
12
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
13
|
+
*/
|
|
14
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
15
|
+
if (k2 === undefined) k2 = k;
|
|
16
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
17
|
+
}) : (function(o, m, k, k2) {
|
|
18
|
+
if (k2 === undefined) k2 = k;
|
|
19
|
+
o[k2] = m[k];
|
|
20
|
+
}));
|
|
21
|
+
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
22
|
+
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
23
|
+
}) : function(o, v) {
|
|
24
|
+
o["default"] = v;
|
|
25
|
+
});
|
|
26
|
+
var __importStar = (this && this.__importStar) || function (mod) {
|
|
27
|
+
if (mod && mod.__esModule) return mod;
|
|
28
|
+
var result = {};
|
|
29
|
+
if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
30
|
+
__setModuleDefault(result, mod);
|
|
31
|
+
return result;
|
|
32
|
+
};
|
|
33
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
34
|
+
exports.VideoFeature = void 0;
|
|
35
|
+
const model = __importStar(require("../model"));
|
|
36
|
+
const common = require("oci-common");
|
|
37
|
+
var VideoFeature;
|
|
38
|
+
(function (VideoFeature) {
|
|
39
|
+
function getJsonObj(obj) {
|
|
40
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
41
|
+
if (obj && "featureType" in obj && obj.featureType) {
|
|
42
|
+
switch (obj.featureType) {
|
|
43
|
+
case "OBJECT_DETECTION":
|
|
44
|
+
return model.VideoObjectDetectionFeature.getJsonObj(jsonObj, true);
|
|
45
|
+
case "FACE_DETECTION":
|
|
46
|
+
return model.VideoFaceDetectionFeature.getJsonObj(jsonObj, true);
|
|
47
|
+
case "TEXT_DETECTION":
|
|
48
|
+
return model.VideoTextDetectionFeature.getJsonObj(jsonObj, true);
|
|
49
|
+
case "OBJECT_TRACKING":
|
|
50
|
+
return model.VideoObjectTrackingFeature.getJsonObj(jsonObj, true);
|
|
51
|
+
case "LABEL_DETECTION":
|
|
52
|
+
return model.VideoLabelDetectionFeature.getJsonObj(jsonObj, true);
|
|
53
|
+
default:
|
|
54
|
+
if (common.LOG.logger)
|
|
55
|
+
common.LOG.logger.info(`Unknown value for: ${obj.featureType}`);
|
|
56
|
+
}
|
|
57
|
+
}
|
|
58
|
+
return jsonObj;
|
|
59
|
+
}
|
|
60
|
+
VideoFeature.getJsonObj = getJsonObj;
|
|
61
|
+
function getDeserializedJsonObj(obj) {
|
|
62
|
+
const jsonObj = Object.assign(Object.assign({}, obj), {});
|
|
63
|
+
if (obj && "featureType" in obj && obj.featureType) {
|
|
64
|
+
switch (obj.featureType) {
|
|
65
|
+
case "OBJECT_DETECTION":
|
|
66
|
+
return model.VideoObjectDetectionFeature.getDeserializedJsonObj(jsonObj, true);
|
|
67
|
+
case "FACE_DETECTION":
|
|
68
|
+
return model.VideoFaceDetectionFeature.getDeserializedJsonObj(jsonObj, true);
|
|
69
|
+
case "TEXT_DETECTION":
|
|
70
|
+
return model.VideoTextDetectionFeature.getDeserializedJsonObj(jsonObj, true);
|
|
71
|
+
case "OBJECT_TRACKING":
|
|
72
|
+
return model.VideoObjectTrackingFeature.getDeserializedJsonObj(jsonObj, true);
|
|
73
|
+
case "LABEL_DETECTION":
|
|
74
|
+
return model.VideoLabelDetectionFeature.getDeserializedJsonObj(jsonObj, true);
|
|
75
|
+
default:
|
|
76
|
+
if (common.LOG.logger)
|
|
77
|
+
common.LOG.logger.info(`Unknown value for: ${obj.featureType}`);
|
|
78
|
+
}
|
|
79
|
+
}
|
|
80
|
+
return jsonObj;
|
|
81
|
+
}
|
|
82
|
+
VideoFeature.getDeserializedJsonObj = getDeserializedJsonObj;
|
|
83
|
+
})(VideoFeature = exports.VideoFeature || (exports.VideoFeature = {}));
|
|
84
|
+
//# sourceMappingURL=video-feature.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"video-feature.js","sourceRoot":"","sources":["../../../../../lib/aivision/lib/model/video-feature.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;GAWG;;;;;;;;;;;;;;;;;;;;;;AAEH,gDAAkC;AAClC,qCAAsC;AAStC,IAAiB,YAAY,CAyE5B;AAzED,WAAiB,YAAY;IAC3B,SAAgB,UAAU,CAAC,GAAiB;QAC1C,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,aAAa,IAAI,GAAG,IAAI,GAAG,CAAC,WAAW,EAAE;YAClD,QAAQ,GAAG,CAAC,WAAW,EAAE;gBACvB,KAAK,kBAAkB;oBACrB,OAAO,KAAK,CAAC,2BAA2B,CAAC,UAAU,CACL,OAAQ,EACpD,IAAI,CACL,CAAC;gBACJ,KAAK,gBAAgB;oBACnB,OAAO,KAAK,CAAC,yBAAyB,CAAC,UAAU,CACL,OAAQ,EAClD,IAAI,CACL,CAAC;gBACJ,KAAK,gBAAgB;oBACnB,OAAO,KAAK,CAAC,yBAAyB,CAAC,UAAU,CACL,OAAQ,EAClD,IAAI,CACL,CAAC;gBACJ,KAAK,iBAAiB;oBACpB,OAAO,KAAK,CAAC,0BAA0B,CAAC,UAAU,CACL,OAAQ,EACnD,IAAI,CACL,CAAC;gBACJ,KAAK,iBAAiB;oBACpB,OAAO,KAAK,CAAC,0BAA0B,CAAC,UAAU,CACL,OAAQ,EACnD,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,WAAW,EAAE,CAAC,CAAC;aAC1F;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAnCe,uBAAU,aAmCzB,CAAA;IACD,SAAgB,sBAAsB,CAAC,GAAiB;QACtD,MAAM,OAAO,mCAAQ,GAAG,GAAK,EAAE,CAAE,CAAC;QAElC,IAAI,GAAG,IAAI,aAAa,IAAI,GAAG,IAAI,GAAG,CAAC,WAAW,EAAE;YAClD,QAAQ,GAAG,CAAC,WAAW,EAAE;gBACvB,KAAK,kBAAkB;oBACrB,OAAO,KAAK,CAAC,2BAA2B,CAAC,sBAAsB,CACjB,OAAQ,EACpD,IAAI,CACL,CAAC;gBACJ,KAAK,gBAAgB;oBACnB,OAAO,KAAK,CAAC,yBAAyB,CAAC,sBAAsB,CACjB,OAAQ,EAClD,IAAI,CACL,CAAC;gBACJ,KAAK,gBAAgB;oBACnB,OAAO,KAAK,CAAC,yBAAyB,CAAC,sBAAsB,CACjB,OAAQ,EAClD,IAAI,CACL,CAAC;gBACJ,KAAK,iBAAiB;oBACpB,OAAO,KAAK,CAAC,0BAA0B,CAAC,sBAAsB,CACjB,OAAQ,EACnD,IAAI,CACL,CAAC;gBACJ,KAAK,iBAAiB;oBACpB,OAAO,KAAK,CAAC,0BAA0B,CAAC,sBAAsB,CACjB,OAAQ,EACnD,IAAI,CACL,CAAC;gBACJ;oBACE,IAAI,MAAM,CAAC,GAAG,CAAC,MAAM;wBAAE,MAAM,CAAC,GAAG,CAAC,MAAM,CAAC,IAAI,CAAC,sBAAsB,GAAG,CAAC,WAAW,EAAE,CAAC,CAAC;aAC1F;SACF;QACD,OAAO,OAAO,CAAC;IACjB,CAAC;IAnCe,mCAAsB,yBAmCrC,CAAA;AACH,CAAC,EAzEgB,YAAY,GAAZ,oBAAY,KAAZ,oBAAY,QAyE5B"}
|
|
@@ -0,0 +1,114 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Vision API
|
|
3
|
+
* Using Vision, you can upload images to detect and classify objects in them. If you have lots of images, you can process them in batch using asynchronous API endpoints. Vision's features are thematically split between Document AI for document-centric images, and Image Analysis for object and scene-based images. Pretrained models and custom models are supported.
|
|
4
|
+
* OpenAPI spec version: 20220125
|
|
5
|
+
*
|
|
6
|
+
*
|
|
7
|
+
* NOTE: This class is auto generated by OracleSDKGenerator.
|
|
8
|
+
* Do not edit the class manually.
|
|
9
|
+
*
|
|
10
|
+
* Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
|
|
11
|
+
* This software is dual-licensed to you under the Universal Permissive License (UPL) 1.0 as shown at https://oss.oracle.com/licenses/upl or Apache License 2.0 as shown at http://www.apache.org/licenses/LICENSE-2.0. You may choose either license.
|
|
12
|
+
*/
|
|
13
|
+
import * as model from "../model";
|
|
14
|
+
/**
|
|
15
|
+
* Job details for a video analysis.
|
|
16
|
+
*/
|
|
17
|
+
export interface VideoJob {
|
|
18
|
+
/**
|
|
19
|
+
* Id of the job.
|
|
20
|
+
*/
|
|
21
|
+
"id": string;
|
|
22
|
+
/**
|
|
23
|
+
* The ocid of the compartment that starts the job.
|
|
24
|
+
*/
|
|
25
|
+
"compartmentId": string;
|
|
26
|
+
/**
|
|
27
|
+
* Video job display name.
|
|
28
|
+
*/
|
|
29
|
+
"displayName"?: string;
|
|
30
|
+
/**
|
|
31
|
+
* a list of document analysis features.
|
|
32
|
+
*/
|
|
33
|
+
"features": Array<model.VideoFeature>;
|
|
34
|
+
"inputLocation"?: model.ObjectListInlineInputLocation;
|
|
35
|
+
/**
|
|
36
|
+
* Job accepted time.
|
|
37
|
+
*/
|
|
38
|
+
"timeAccepted": Date;
|
|
39
|
+
/**
|
|
40
|
+
* Job started time.
|
|
41
|
+
*/
|
|
42
|
+
"timeStarted"?: Date;
|
|
43
|
+
/**
|
|
44
|
+
* Job finished time.
|
|
45
|
+
*/
|
|
46
|
+
"timeFinished"?: Date;
|
|
47
|
+
/**
|
|
48
|
+
* How much progress the operation has made, vs the total amount of work that must be performed. Note: Numbers greater than Number.MAX_SAFE_INTEGER will result in rounding issues.
|
|
49
|
+
*/
|
|
50
|
+
"percentComplete"?: number;
|
|
51
|
+
"outputLocation": model.OutputLocation;
|
|
52
|
+
/**
|
|
53
|
+
* The current state of the batch document job.
|
|
54
|
+
*/
|
|
55
|
+
"lifecycleState": VideoJob.LifecycleState;
|
|
56
|
+
/**
|
|
57
|
+
* Detailed status of FAILED state.
|
|
58
|
+
*/
|
|
59
|
+
"lifecycleDetails"?: VideoJob.LifecycleDetails;
|
|
60
|
+
/**
|
|
61
|
+
* A simple key-value pair that is applied without any predefined name, type, or scope. It exists for cross-compatibility only.
|
|
62
|
+
* For example: {@code {\"bar-key\": \"value\"}}
|
|
63
|
+
*
|
|
64
|
+
*/
|
|
65
|
+
"freeformTags"?: {
|
|
66
|
+
[key: string]: string;
|
|
67
|
+
};
|
|
68
|
+
/**
|
|
69
|
+
* Defined tags for this resource. Each key is predefined and scoped to a namespace.
|
|
70
|
+
* For example: {@code {\"foo-namespace\": {\"bar-key\": \"value\"}}}
|
|
71
|
+
*
|
|
72
|
+
*/
|
|
73
|
+
"definedTags"?: {
|
|
74
|
+
[key: string]: {
|
|
75
|
+
[key: string]: any;
|
|
76
|
+
};
|
|
77
|
+
};
|
|
78
|
+
/**
|
|
79
|
+
* Usage of system tag keys. These predefined keys are scoped to namespaces.
|
|
80
|
+
* For example: {@code {\"orcl-cloud\": {\"free-tier-retained\": \"true\"}}}
|
|
81
|
+
*
|
|
82
|
+
*/
|
|
83
|
+
"systemTags"?: {
|
|
84
|
+
[key: string]: {
|
|
85
|
+
[key: string]: any;
|
|
86
|
+
};
|
|
87
|
+
};
|
|
88
|
+
}
|
|
89
|
+
export declare namespace VideoJob {
|
|
90
|
+
enum LifecycleState {
|
|
91
|
+
Succeeded = "SUCCEEDED",
|
|
92
|
+
Failed = "FAILED",
|
|
93
|
+
Accepted = "ACCEPTED",
|
|
94
|
+
Canceled = "CANCELED",
|
|
95
|
+
InProgress = "IN_PROGRESS",
|
|
96
|
+
Canceling = "CANCELING",
|
|
97
|
+
/**
|
|
98
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
99
|
+
* version of the SDK.
|
|
100
|
+
*/
|
|
101
|
+
UnknownValue = "UNKNOWN_VALUE"
|
|
102
|
+
}
|
|
103
|
+
enum LifecycleDetails {
|
|
104
|
+
PartiallySucceeded = "PARTIALLY_SUCCEEDED",
|
|
105
|
+
CompletelyFailed = "COMPLETELY_FAILED",
|
|
106
|
+
/**
|
|
107
|
+
* This value is used if a service returns a value for this enum that is not recognized by this
|
|
108
|
+
* version of the SDK.
|
|
109
|
+
*/
|
|
110
|
+
UnknownValue = "UNKNOWN_VALUE"
|
|
111
|
+
}
|
|
112
|
+
function getJsonObj(obj: VideoJob): object;
|
|
113
|
+
function getDeserializedJsonObj(obj: VideoJob): object;
|
|
114
|
+
}
|