web-speech-cognitive-services 7.1.4-master.151bc9b → 8.0.0-main.15b930d
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/web-speech-cognitive-services.d.mts +329 -0
- package/dist/web-speech-cognitive-services.d.ts +329 -0
- package/dist/web-speech-cognitive-services.development.js +25746 -0
- package/dist/web-speech-cognitive-services.development.js.map +1 -0
- package/dist/web-speech-cognitive-services.js +1409 -0
- package/dist/web-speech-cognitive-services.js.map +1 -0
- package/dist/web-speech-cognitive-services.mjs +1374 -0
- package/dist/web-speech-cognitive-services.mjs.map +1 -0
- package/dist/web-speech-cognitive-services.production.min.js +31 -0
- package/dist/web-speech-cognitive-services.production.min.js.map +1 -0
- package/package.json +67 -47
- package/CHANGELOG.md +0 -372
- package/lib/BingSpeech/SpeechToText/SpeechGrammarList.js +0 -94
- package/lib/BingSpeech/SpeechToText/SpeechGrammarList.js.map +0 -1
- package/lib/BingSpeech/SpeechToText/createSpeechRecognitionPonyfill.js +0 -483
- package/lib/BingSpeech/SpeechToText/createSpeechRecognitionPonyfill.js.map +0 -1
- package/lib/BingSpeech/SpeechToText.js +0 -14
- package/lib/BingSpeech/SpeechToText.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/AudioContextConsumer.js +0 -122
- package/lib/BingSpeech/TextToSpeech/AudioContextConsumer.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/AudioContextQueue.js +0 -104
- package/lib/BingSpeech/TextToSpeech/AudioContextQueue.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/SpeechSynthesisUtterance.js +0 -264
- package/lib/BingSpeech/TextToSpeech/SpeechSynthesisUtterance.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/SpeechSynthesisVoice.js +0 -61
- package/lib/BingSpeech/TextToSpeech/SpeechSynthesisVoice.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/buildSSML.js +0 -32
- package/lib/BingSpeech/TextToSpeech/buildSSML.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/createSpeechSynthesisPonyfill.js +0 -220
- package/lib/BingSpeech/TextToSpeech/createSpeechSynthesisPonyfill.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/fetchSpeechData.js +0 -74
- package/lib/BingSpeech/TextToSpeech/fetchSpeechData.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/fetchVoices.js +0 -335
- package/lib/BingSpeech/TextToSpeech/fetchVoices.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/isSSML.js +0 -13
- package/lib/BingSpeech/TextToSpeech/isSSML.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech/subscribeEvent.js +0 -14
- package/lib/BingSpeech/TextToSpeech/subscribeEvent.js.map +0 -1
- package/lib/BingSpeech/TextToSpeech.js +0 -14
- package/lib/BingSpeech/TextToSpeech.js.map +0 -1
- package/lib/BingSpeech/Util/DOMEventEmitter.js +0 -61
- package/lib/BingSpeech/Util/DOMEventEmitter.js.map +0 -1
- package/lib/BingSpeech/Util/createFetchTokenUsingSubscriptionKey.js +0 -41
- package/lib/BingSpeech/Util/createFetchTokenUsingSubscriptionKey.js.map +0 -1
- package/lib/BingSpeech/fetchAuthorizationToken.js +0 -57
- package/lib/BingSpeech/fetchAuthorizationToken.js.map +0 -1
- package/lib/BingSpeech/index.js +0 -84
- package/lib/BingSpeech/index.js.map +0 -1
- package/lib/SpeechServices/SpeechSDK.js +0 -19
- package/lib/SpeechServices/SpeechSDK.js.map +0 -1
- package/lib/SpeechServices/SpeechToText/SpeechGrammarList.js +0 -45
- package/lib/SpeechServices/SpeechToText/SpeechGrammarList.js.map +0 -1
- package/lib/SpeechServices/SpeechToText/cognitiveServiceEventResultToWebSpeechRecognitionResultList.js +0 -56
- package/lib/SpeechServices/SpeechToText/cognitiveServiceEventResultToWebSpeechRecognitionResultList.js.map +0 -1
- package/lib/SpeechServices/SpeechToText/createSpeechRecognitionPonyfill.js +0 -984
- package/lib/SpeechServices/SpeechToText/createSpeechRecognitionPonyfill.js.map +0 -1
- package/lib/SpeechServices/SpeechToText.js +0 -24
- package/lib/SpeechServices/SpeechToText.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/AudioContextConsumer.js +0 -92
- package/lib/SpeechServices/TextToSpeech/AudioContextConsumer.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/AudioContextQueue.js +0 -111
- package/lib/SpeechServices/TextToSpeech/AudioContextQueue.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/SpeechSynthesisEvent.js +0 -40
- package/lib/SpeechServices/TextToSpeech/SpeechSynthesisEvent.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/SpeechSynthesisUtterance.js +0 -283
- package/lib/SpeechServices/TextToSpeech/SpeechSynthesisUtterance.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/SpeechSynthesisVoice.js +0 -63
- package/lib/SpeechServices/TextToSpeech/SpeechSynthesisVoice.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/buildSSML.js +0 -32
- package/lib/SpeechServices/TextToSpeech/buildSSML.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/createSpeechSynthesisPonyfill.js +0 -282
- package/lib/SpeechServices/TextToSpeech/createSpeechSynthesisPonyfill.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/fetchCustomVoices.js +0 -110
- package/lib/SpeechServices/TextToSpeech/fetchCustomVoices.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/fetchSpeechData.js +0 -127
- package/lib/SpeechServices/TextToSpeech/fetchSpeechData.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/fetchVoices.js +0 -87
- package/lib/SpeechServices/TextToSpeech/fetchVoices.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/isSSML.js +0 -13
- package/lib/SpeechServices/TextToSpeech/isSSML.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech/subscribeEvent.js +0 -14
- package/lib/SpeechServices/TextToSpeech/subscribeEvent.js.map +0 -1
- package/lib/SpeechServices/TextToSpeech.js +0 -14
- package/lib/SpeechServices/TextToSpeech.js.map +0 -1
- package/lib/SpeechServices/fetchAuthorizationToken.js +0 -58
- package/lib/SpeechServices/fetchAuthorizationToken.js.map +0 -1
- package/lib/SpeechServices/patchOptions.js +0 -213
- package/lib/SpeechServices/patchOptions.js.map +0 -1
- package/lib/SpeechServices/resolveFunctionOrReturnValue.js +0 -11
- package/lib/SpeechServices/resolveFunctionOrReturnValue.js.map +0 -1
- package/lib/SpeechServices.js +0 -73
- package/lib/SpeechServices.js.map +0 -1
- package/lib/Util/arrayToMap.js +0 -28
- package/lib/Util/arrayToMap.js.map +0 -1
- package/lib/Util/createPromiseQueue.js +0 -40
- package/lib/Util/createPromiseQueue.js.map +0 -1
- package/lib/index.js +0 -14
- package/lib/index.js.map +0 -1
- package/umd/web-speech-cognitive-services.development.js +0 -4740
- package/umd/web-speech-cognitive-services.production.min.js +0 -2
|
@@ -1,220 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
|
|
3
|
-
var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
|
|
4
|
-
|
|
5
|
-
Object.defineProperty(exports, "__esModule", {
|
|
6
|
-
value: true
|
|
7
|
-
});
|
|
8
|
-
exports.default = void 0;
|
|
9
|
-
|
|
10
|
-
var _regenerator = _interopRequireDefault(require("@babel/runtime/regenerator"));
|
|
11
|
-
|
|
12
|
-
var _asyncToGenerator2 = _interopRequireDefault(require("@babel/runtime/helpers/asyncToGenerator"));
|
|
13
|
-
|
|
14
|
-
var _classCallCheck2 = _interopRequireDefault(require("@babel/runtime/helpers/classCallCheck"));
|
|
15
|
-
|
|
16
|
-
var _createClass2 = _interopRequireDefault(require("@babel/runtime/helpers/createClass"));
|
|
17
|
-
|
|
18
|
-
var _inherits2 = _interopRequireDefault(require("@babel/runtime/helpers/inherits"));
|
|
19
|
-
|
|
20
|
-
var _possibleConstructorReturn2 = _interopRequireDefault(require("@babel/runtime/helpers/possibleConstructorReturn"));
|
|
21
|
-
|
|
22
|
-
var _getPrototypeOf2 = _interopRequireDefault(require("@babel/runtime/helpers/getPrototypeOf"));
|
|
23
|
-
|
|
24
|
-
var _memoizeOne = _interopRequireDefault(require("memoize-one"));
|
|
25
|
-
|
|
26
|
-
var _AudioContextQueue = _interopRequireDefault(require("./AudioContextQueue"));
|
|
27
|
-
|
|
28
|
-
var _DOMEventEmitter2 = _interopRequireDefault(require("../Util/DOMEventEmitter"));
|
|
29
|
-
|
|
30
|
-
var _fetchAuthorizationToken = _interopRequireDefault(require("../fetchAuthorizationToken"));
|
|
31
|
-
|
|
32
|
-
var _fetchVoices = _interopRequireDefault(require("./fetchVoices"));
|
|
33
|
-
|
|
34
|
-
var _SpeechSynthesisUtterance = _interopRequireDefault(require("./SpeechSynthesisUtterance"));
|
|
35
|
-
|
|
36
|
-
function _createSuper(Derived) { var hasNativeReflectConstruct = _isNativeReflectConstruct(); return function _createSuperInternal() { var Super = (0, _getPrototypeOf2.default)(Derived), result; if (hasNativeReflectConstruct) { var NewTarget = (0, _getPrototypeOf2.default)(this).constructor; result = Reflect.construct(Super, arguments, NewTarget); } else { result = Super.apply(this, arguments); } return (0, _possibleConstructorReturn2.default)(this, result); }; }
|
|
37
|
-
|
|
38
|
-
function _isNativeReflectConstruct() { if (typeof Reflect === "undefined" || !Reflect.construct) return false; if (Reflect.construct.sham) return false; if (typeof Proxy === "function") return true; try { Boolean.prototype.valueOf.call(Reflect.construct(Boolean, [], function () {})); return true; } catch (e) { return false; } }
|
|
39
|
-
|
|
40
|
-
// Supported output format can be found at https://docs.microsoft.com/en-us/azure/cognitive-services/Speech/API-Reference-REST/BingVoiceOutput#Subscription
|
|
41
|
-
var DEFAULT_OUTPUT_FORMAT = 'audio-16khz-128kbitrate-mono-mp3';
|
|
42
|
-
var TOKEN_EXPIRATION = 600000;
|
|
43
|
-
var TOKEN_EARLY_RENEWAL = 60000;
|
|
44
|
-
|
|
45
|
-
var _default = function _default(_ref) {
|
|
46
|
-
var authorizationToken = _ref.authorizationToken,
|
|
47
|
-
_ref$ponyfill = _ref.ponyfill,
|
|
48
|
-
ponyfill = _ref$ponyfill === void 0 ? {
|
|
49
|
-
AudioContext: window.AudioContext || window.webkitAudioContext
|
|
50
|
-
} : _ref$ponyfill,
|
|
51
|
-
subscriptionKey = _ref.subscriptionKey;
|
|
52
|
-
|
|
53
|
-
if (!authorizationToken && !subscriptionKey) {
|
|
54
|
-
console.warn('Either authorizationToken or subscriptionKey must be specified');
|
|
55
|
-
return {};
|
|
56
|
-
} else if (!ponyfill.AudioContext) {
|
|
57
|
-
console.warn('This browser does not support Web Audio and it will not work with Cognitive Services Speech Services.');
|
|
58
|
-
return {};
|
|
59
|
-
}
|
|
60
|
-
|
|
61
|
-
var fetchMemoizedAuthorizationToken = (0, _memoizeOne.default)(function (_ref2) {
|
|
62
|
-
var subscriptionKey = _ref2.subscriptionKey;
|
|
63
|
-
return (0, _fetchAuthorizationToken.default)(subscriptionKey);
|
|
64
|
-
}, function (arg, prevArg) {
|
|
65
|
-
return arg.subscriptionKey === prevArg.subscriptionKey && arg.now - prevArg.now < TOKEN_EXPIRATION - TOKEN_EARLY_RENEWAL;
|
|
66
|
-
});
|
|
67
|
-
|
|
68
|
-
var SpeechSynthesis = /*#__PURE__*/function (_DOMEventEmitter) {
|
|
69
|
-
(0, _inherits2.default)(SpeechSynthesis, _DOMEventEmitter);
|
|
70
|
-
|
|
71
|
-
var _super = _createSuper(SpeechSynthesis);
|
|
72
|
-
|
|
73
|
-
function SpeechSynthesis() {
|
|
74
|
-
var _this;
|
|
75
|
-
|
|
76
|
-
(0, _classCallCheck2.default)(this, SpeechSynthesis);
|
|
77
|
-
_this = _super.call(this, ['voiceschanged']);
|
|
78
|
-
_this.outputFormat = DEFAULT_OUTPUT_FORMAT;
|
|
79
|
-
_this.queue = new _AudioContextQueue.default(ponyfill);
|
|
80
|
-
return _this;
|
|
81
|
-
}
|
|
82
|
-
|
|
83
|
-
(0, _createClass2.default)(SpeechSynthesis, [{
|
|
84
|
-
key: "cancel",
|
|
85
|
-
value: function cancel() {
|
|
86
|
-
this.queue.stop();
|
|
87
|
-
}
|
|
88
|
-
}, {
|
|
89
|
-
key: "getVoices",
|
|
90
|
-
value: function getVoices() {
|
|
91
|
-
return (0, _fetchVoices.default)();
|
|
92
|
-
}
|
|
93
|
-
}, {
|
|
94
|
-
key: "pause",
|
|
95
|
-
value: function pause() {
|
|
96
|
-
this.queue.pause();
|
|
97
|
-
}
|
|
98
|
-
}, {
|
|
99
|
-
key: "resume",
|
|
100
|
-
value: function resume() {
|
|
101
|
-
this.queue.resume();
|
|
102
|
-
}
|
|
103
|
-
}, {
|
|
104
|
-
key: "speak",
|
|
105
|
-
value: function () {
|
|
106
|
-
var _speak = (0, _asyncToGenerator2.default)( /*#__PURE__*/_regenerator.default.mark(function _callee2(utterance) {
|
|
107
|
-
var _this2 = this;
|
|
108
|
-
|
|
109
|
-
return _regenerator.default.wrap(function _callee2$(_context2) {
|
|
110
|
-
while (1) {
|
|
111
|
-
switch (_context2.prev = _context2.next) {
|
|
112
|
-
case 0:
|
|
113
|
-
if (utterance instanceof _SpeechSynthesisUtterance.default) {
|
|
114
|
-
_context2.next = 2;
|
|
115
|
-
break;
|
|
116
|
-
}
|
|
117
|
-
|
|
118
|
-
throw new Error('invalid utterance');
|
|
119
|
-
|
|
120
|
-
case 2:
|
|
121
|
-
return _context2.abrupt("return", new Promise( /*#__PURE__*/function () {
|
|
122
|
-
var _ref3 = (0, _asyncToGenerator2.default)( /*#__PURE__*/_regenerator.default.mark(function _callee(resolve, reject) {
|
|
123
|
-
return _regenerator.default.wrap(function _callee$(_context) {
|
|
124
|
-
while (1) {
|
|
125
|
-
switch (_context.prev = _context.next) {
|
|
126
|
-
case 0:
|
|
127
|
-
utterance.addEventListener('end', resolve);
|
|
128
|
-
utterance.addEventListener('error', reject);
|
|
129
|
-
|
|
130
|
-
if (!(typeof authorizationToken === 'function')) {
|
|
131
|
-
_context.next = 8;
|
|
132
|
-
break;
|
|
133
|
-
}
|
|
134
|
-
|
|
135
|
-
_context.next = 5;
|
|
136
|
-
return authorizationToken();
|
|
137
|
-
|
|
138
|
-
case 5:
|
|
139
|
-
_context.t0 = _context.sent;
|
|
140
|
-
_context.next = 18;
|
|
141
|
-
break;
|
|
142
|
-
|
|
143
|
-
case 8:
|
|
144
|
-
if (!authorizationToken) {
|
|
145
|
-
_context.next = 14;
|
|
146
|
-
break;
|
|
147
|
-
}
|
|
148
|
-
|
|
149
|
-
_context.next = 11;
|
|
150
|
-
return authorizationToken;
|
|
151
|
-
|
|
152
|
-
case 11:
|
|
153
|
-
_context.t1 = _context.sent;
|
|
154
|
-
_context.next = 17;
|
|
155
|
-
break;
|
|
156
|
-
|
|
157
|
-
case 14:
|
|
158
|
-
_context.next = 16;
|
|
159
|
-
return fetchMemoizedAuthorizationToken({
|
|
160
|
-
now: Date.now,
|
|
161
|
-
subscriptionKey: subscriptionKey
|
|
162
|
-
});
|
|
163
|
-
|
|
164
|
-
case 16:
|
|
165
|
-
_context.t1 = _context.sent;
|
|
166
|
-
|
|
167
|
-
case 17:
|
|
168
|
-
_context.t0 = _context.t1;
|
|
169
|
-
|
|
170
|
-
case 18:
|
|
171
|
-
utterance.authorizationToken = _context.t0;
|
|
172
|
-
utterance.outputFormat = _this2.outputFormat;
|
|
173
|
-
utterance.preload();
|
|
174
|
-
|
|
175
|
-
_this2.queue.push(utterance);
|
|
176
|
-
|
|
177
|
-
case 22:
|
|
178
|
-
case "end":
|
|
179
|
-
return _context.stop();
|
|
180
|
-
}
|
|
181
|
-
}
|
|
182
|
-
}, _callee);
|
|
183
|
-
}));
|
|
184
|
-
|
|
185
|
-
return function (_x2, _x3) {
|
|
186
|
-
return _ref3.apply(this, arguments);
|
|
187
|
-
};
|
|
188
|
-
}()));
|
|
189
|
-
|
|
190
|
-
case 3:
|
|
191
|
-
case "end":
|
|
192
|
-
return _context2.stop();
|
|
193
|
-
}
|
|
194
|
-
}
|
|
195
|
-
}, _callee2);
|
|
196
|
-
}));
|
|
197
|
-
|
|
198
|
-
function speak(_x) {
|
|
199
|
-
return _speak.apply(this, arguments);
|
|
200
|
-
}
|
|
201
|
-
|
|
202
|
-
return speak;
|
|
203
|
-
}()
|
|
204
|
-
}, {
|
|
205
|
-
key: "speaking",
|
|
206
|
-
get: function get() {
|
|
207
|
-
return this.queue.speaking;
|
|
208
|
-
}
|
|
209
|
-
}]);
|
|
210
|
-
return SpeechSynthesis;
|
|
211
|
-
}(_DOMEventEmitter2.default);
|
|
212
|
-
|
|
213
|
-
return {
|
|
214
|
-
speechSynthesis: new SpeechSynthesis(),
|
|
215
|
-
SpeechSynthesisUtterance: _SpeechSynthesisUtterance.default
|
|
216
|
-
};
|
|
217
|
-
};
|
|
218
|
-
|
|
219
|
-
exports.default = _default;
|
|
220
|
-
//# sourceMappingURL=createSpeechSynthesisPonyfill.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"createSpeechSynthesisPonyfill.js","names":["DEFAULT_OUTPUT_FORMAT","TOKEN_EXPIRATION","TOKEN_EARLY_RENEWAL","authorizationToken","ponyfill","AudioContext","window","webkitAudioContext","subscriptionKey","console","warn","fetchMemoizedAuthorizationToken","memoize","fetchAuthorizationToken","arg","prevArg","now","SpeechSynthesis","outputFormat","queue","AudioContextQueue","stop","fetchVoices","pause","resume","utterance","SpeechSynthesisUtterance","Error","Promise","resolve","reject","addEventListener","Date","preload","push","speaking","DOMEventEmitter","speechSynthesis"],"sources":["../../../src/BingSpeech/TextToSpeech/createSpeechSynthesisPonyfill.js"],"sourcesContent":["import memoize from 'memoize-one';\n\nimport AudioContextQueue from './AudioContextQueue';\nimport DOMEventEmitter from '../Util/DOMEventEmitter';\nimport fetchAuthorizationToken from '../fetchAuthorizationToken';\nimport fetchVoices from './fetchVoices';\nimport SpeechSynthesisUtterance from './SpeechSynthesisUtterance';\n\n// Supported output format can be found at https://docs.microsoft.com/en-us/azure/cognitive-services/Speech/API-Reference-REST/BingVoiceOutput#Subscription\nconst DEFAULT_OUTPUT_FORMAT = 'audio-16khz-128kbitrate-mono-mp3';\n\nconst TOKEN_EXPIRATION = 600000;\nconst TOKEN_EARLY_RENEWAL = 60000;\n\nexport default ({\n authorizationToken,\n ponyfill = {\n AudioContext: window.AudioContext || window.webkitAudioContext\n },\n subscriptionKey\n}) => {\n if (!authorizationToken && !subscriptionKey) {\n console.warn('Either authorizationToken or subscriptionKey must be specified');\n\n return {};\n } else if (!ponyfill.AudioContext) {\n console.warn('This browser does not support Web Audio and it will not work with Cognitive Services Speech Services.');\n\n return {};\n }\n\n const fetchMemoizedAuthorizationToken = memoize(\n ({ subscriptionKey }) => fetchAuthorizationToken(subscriptionKey),\n (arg, prevArg) => (\n arg.subscriptionKey === prevArg.subscriptionKey\n && arg.now - prevArg.now < TOKEN_EXPIRATION - TOKEN_EARLY_RENEWAL\n )\n );\n\n class SpeechSynthesis extends DOMEventEmitter {\n constructor() {\n super(['voiceschanged']);\n\n this.outputFormat = DEFAULT_OUTPUT_FORMAT;\n this.queue = new AudioContextQueue(ponyfill);\n }\n\n cancel() {\n this.queue.stop();\n }\n\n getVoices() {\n return fetchVoices();\n }\n\n pause() {\n this.queue.pause();\n }\n\n resume() {\n this.queue.resume();\n }\n\n async speak(utterance) {\n if (!(utterance instanceof SpeechSynthesisUtterance)) {\n throw new Error('invalid utterance');\n }\n\n return new Promise(async (resolve, reject) => {\n utterance.addEventListener('end', resolve);\n utterance.addEventListener('error', reject);\n\n utterance.authorizationToken =\n typeof authorizationToken === 'function' ?\n await authorizationToken()\n : authorizationToken ?\n await authorizationToken\n : await fetchMemoizedAuthorizationToken({\n now: Date.now,\n subscriptionKey\n });\n utterance.outputFormat = this.outputFormat;\n utterance.preload();\n\n this.queue.push(utterance);\n });\n }\n\n get speaking() {\n return this.queue.speaking;\n }\n }\n\n return {\n speechSynthesis: new SpeechSynthesis(),\n SpeechSynthesisUtterance\n };\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;AAAA;;AAEA;;AACA;;AACA;;AACA;;AACA;;;;;;AAEA;AACA,IAAMA,qBAAqB,GAAG,kCAA9B;AAEA,IAAMC,gBAAgB,GAAG,MAAzB;AACA,IAAMC,mBAAmB,GAAG,KAA5B;;eAEe,wBAMT;EAAA,IALJC,kBAKI,QALJA,kBAKI;EAAA,yBAJJC,QAII;EAAA,IAJJA,QAII,8BAJO;IACTC,YAAY,EAAEC,MAAM,CAACD,YAAP,IAAuBC,MAAM,CAACC;EADnC,CAIP;EAAA,IADJC,eACI,QADJA,eACI;;EACJ,IAAI,CAACL,kBAAD,IAAuB,CAACK,eAA5B,EAA6C;IAC3CC,OAAO,CAACC,IAAR,CAAa,gEAAb;IAEA,OAAO,EAAP;EACD,CAJD,MAIO,IAAI,CAACN,QAAQ,CAACC,YAAd,EAA4B;IACjCI,OAAO,CAACC,IAAR,CAAa,uGAAb;IAEA,OAAO,EAAP;EACD;;EAED,IAAMC,+BAA+B,GAAG,IAAAC,mBAAA,EACtC;IAAA,IAAGJ,eAAH,SAAGA,eAAH;IAAA,OAAyB,IAAAK,gCAAA,EAAwBL,eAAxB,CAAzB;EAAA,CADsC,EAEtC,UAACM,GAAD,EAAMC,OAAN;IAAA,OACED,GAAG,CAACN,eAAJ,KAAwBO,OAAO,CAACP,eAAhC,IACGM,GAAG,CAACE,GAAJ,GAAUD,OAAO,CAACC,GAAlB,GAAwBf,gBAAgB,GAAGC,mBAFhD;EAAA,CAFsC,CAAxC;;EAXI,IAmBEe,eAnBF;IAAA;;IAAA;;IAoBF,2BAAc;MAAA;;MAAA;MACZ,0BAAM,CAAC,eAAD,CAAN;MAEA,MAAKC,YAAL,GAAoBlB,qBAApB;MACA,MAAKmB,KAAL,GAAa,IAAIC,0BAAJ,CAAsBhB,QAAtB,CAAb;MAJY;IAKb;;IAzBC;MAAA;MAAA,OA2BF,kBAAS;QACP,KAAKe,KAAL,CAAWE,IAAX;MACD;IA7BC;MAAA;MAAA,OA+BF,qBAAY;QACV,OAAO,IAAAC,oBAAA,GAAP;MACD;IAjCC;MAAA;MAAA,OAmCF,iBAAQ;QACN,KAAKH,KAAL,CAAWI,KAAX;MACD;IArCC;MAAA;MAAA,OAuCF,kBAAS;QACP,KAAKJ,KAAL,CAAWK,MAAX;MACD;IAzCC;MAAA;MAAA;QAAA,qFA2CF,kBAAYC,SAAZ;UAAA;;UAAA;YAAA;cAAA;gBAAA;kBAAA,IACQA,SAAS,YAAYC,iCAD7B;oBAAA;oBAAA;kBAAA;;kBAAA,MAEU,IAAIC,KAAJ,CAAU,mBAAV,CAFV;;gBAAA;kBAAA,kCAKS,IAAIC,OAAJ;oBAAA,oFAAY,iBAAOC,OAAP,EAAgBC,MAAhB;sBAAA;wBAAA;0BAAA;4BAAA;8BACjBL,SAAS,CAACM,gBAAV,CAA2B,KAA3B,EAAkCF,OAAlC;8BACAJ,SAAS,CAACM,gBAAV,CAA2B,OAA3B,EAAoCD,MAApC;;8BAFiB,MAKf,OAAO3B,kBAAP,KAA8B,UALf;gCAAA;gCAAA;8BAAA;;8BAAA;8BAAA,OAMPA,kBAAkB,EANX;;4BAAA;8BAAA;8BAAA;8BAAA;;4BAAA;8BAAA,KAObA,kBAPa;gCAAA;gCAAA;8BAAA;;8BAAA;8BAAA,OAQPA,kBARO;;4BAAA;8BAAA;8BAAA;8BAAA;;4BAAA;8BAAA;8BAAA,OASPQ,+BAA+B,CAAC;gCACtCK,GAAG,EAAEgB,IAAI,CAAChB,GAD4B;gCAEtCR,eAAe,EAAfA;8BAFsC,CAAD,CATxB;;4BAAA;8BAAA;;4BAAA;8BAAA;;4BAAA;8BAIjBiB,SAAS,CAACtB,kBAJO;8BAajBsB,SAAS,CAACP,YAAV,GAAyB,MAAI,CAACA,YAA9B;8BACAO,SAAS,CAACQ,OAAV;;8BAEA,MAAI,CAACd,KAAL,CAAWe,IAAX,CAAgBT,SAAhB;;4BAhBiB;4BAAA;8BAAA;0BAAA;wBAAA;sBAAA;oBAAA,CAAZ;;oBAAA;sBAAA;oBAAA;kBAAA,IALT;;gBAAA;gBAAA;kBAAA;cAAA;YAAA;UAAA;QAAA,CA3CE;;QAAA;UAAA;QAAA;;QAAA;MAAA;IAAA;MAAA;MAAA,KAoEF,eAAe;QACb,OAAO,KAAKN,KAAL,CAAWgB,QAAlB;MACD;IAtEC;IAAA;EAAA,EAmB0BC,yBAnB1B;;EAyEJ,OAAO;IACLC,eAAe,EAAE,IAAIpB,eAAJ,EADZ;IAELS,wBAAwB,EAAxBA;EAFK,CAAP;AAID,C"}
|
|
@@ -1,74 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
|
|
3
|
-
var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
|
|
4
|
-
|
|
5
|
-
Object.defineProperty(exports, "__esModule", {
|
|
6
|
-
value: true
|
|
7
|
-
});
|
|
8
|
-
exports.default = fetchSpeechData;
|
|
9
|
-
|
|
10
|
-
var _regenerator = _interopRequireDefault(require("@babel/runtime/regenerator"));
|
|
11
|
-
|
|
12
|
-
var _asyncToGenerator2 = _interopRequireDefault(require("@babel/runtime/helpers/asyncToGenerator"));
|
|
13
|
-
|
|
14
|
-
var _buildSSML = _interopRequireDefault(require("./buildSSML"));
|
|
15
|
-
|
|
16
|
-
var DEFAULT_LANGUAGE = 'en-US';
|
|
17
|
-
var DEFAULT_VOICE = 'Microsoft Server Speech Text to Speech Voice (en-US, JessaRUS)';
|
|
18
|
-
var SYNTHESIS_URL = 'https://speech.platform.bing.com/synthesize';
|
|
19
|
-
|
|
20
|
-
function fetchSpeechData(_x) {
|
|
21
|
-
return _fetchSpeechData.apply(this, arguments);
|
|
22
|
-
}
|
|
23
|
-
|
|
24
|
-
function _fetchSpeechData() {
|
|
25
|
-
_fetchSpeechData = (0, _asyncToGenerator2.default)( /*#__PURE__*/_regenerator.default.mark(function _callee(_ref) {
|
|
26
|
-
var authorizationToken, _ref$lang, lang, outputFormat, pitch, rate, text, _ref$voice, voice, volume, ssml, res;
|
|
27
|
-
|
|
28
|
-
return _regenerator.default.wrap(function _callee$(_context) {
|
|
29
|
-
while (1) {
|
|
30
|
-
switch (_context.prev = _context.next) {
|
|
31
|
-
case 0:
|
|
32
|
-
authorizationToken = _ref.authorizationToken, _ref$lang = _ref.lang, lang = _ref$lang === void 0 ? DEFAULT_LANGUAGE : _ref$lang, outputFormat = _ref.outputFormat, pitch = _ref.pitch, rate = _ref.rate, text = _ref.text, _ref$voice = _ref.voice, voice = _ref$voice === void 0 ? DEFAULT_VOICE : _ref$voice, volume = _ref.volume;
|
|
33
|
-
ssml = (0, _buildSSML.default)({
|
|
34
|
-
lang: lang,
|
|
35
|
-
pitch: pitch,
|
|
36
|
-
rate: rate,
|
|
37
|
-
text: text,
|
|
38
|
-
voice: voice,
|
|
39
|
-
volume: volume
|
|
40
|
-
});
|
|
41
|
-
_context.next = 4;
|
|
42
|
-
return fetch(SYNTHESIS_URL, {
|
|
43
|
-
headers: {
|
|
44
|
-
Authorization: "Bearer ".concat(authorizationToken),
|
|
45
|
-
'Content-Type': 'application/ssml+xml',
|
|
46
|
-
'X-Microsoft-OutputFormat': outputFormat
|
|
47
|
-
},
|
|
48
|
-
method: 'POST',
|
|
49
|
-
body: ssml
|
|
50
|
-
});
|
|
51
|
-
|
|
52
|
-
case 4:
|
|
53
|
-
res = _context.sent;
|
|
54
|
-
|
|
55
|
-
if (!(res.status !== 200)) {
|
|
56
|
-
_context.next = 7;
|
|
57
|
-
break;
|
|
58
|
-
}
|
|
59
|
-
|
|
60
|
-
throw new Error("Failed to syntheis speech, server returned ".concat(res.status));
|
|
61
|
-
|
|
62
|
-
case 7:
|
|
63
|
-
return _context.abrupt("return", res.arrayBuffer());
|
|
64
|
-
|
|
65
|
-
case 8:
|
|
66
|
-
case "end":
|
|
67
|
-
return _context.stop();
|
|
68
|
-
}
|
|
69
|
-
}
|
|
70
|
-
}, _callee);
|
|
71
|
-
}));
|
|
72
|
-
return _fetchSpeechData.apply(this, arguments);
|
|
73
|
-
}
|
|
74
|
-
//# sourceMappingURL=fetchSpeechData.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"fetchSpeechData.js","names":["DEFAULT_LANGUAGE","DEFAULT_VOICE","SYNTHESIS_URL","fetchSpeechData","authorizationToken","lang","outputFormat","pitch","rate","text","voice","volume","ssml","buildSSML","fetch","headers","Authorization","method","body","res","status","Error","arrayBuffer"],"sources":["../../../src/BingSpeech/TextToSpeech/fetchSpeechData.js"],"sourcesContent":["import buildSSML from './buildSSML';\n\nconst DEFAULT_LANGUAGE = 'en-US';\nconst DEFAULT_VOICE = 'Microsoft Server Speech Text to Speech Voice (en-US, JessaRUS)'\nconst SYNTHESIS_URL = 'https://speech.platform.bing.com/synthesize';\n\nexport default async function fetchSpeechData({\n authorizationToken,\n lang = DEFAULT_LANGUAGE,\n outputFormat,\n pitch,\n rate,\n text,\n voice = DEFAULT_VOICE,\n volume\n}) {\n const ssml = buildSSML({ lang, pitch, rate, text, voice, volume });\n\n const res = await fetch(SYNTHESIS_URL, {\n headers: {\n Authorization: `Bearer ${ authorizationToken }`,\n 'Content-Type': 'application/ssml+xml',\n 'X-Microsoft-OutputFormat': outputFormat\n },\n method: 'POST',\n body: ssml\n });\n\n if (res.status !== 200) {\n throw new Error(`Failed to syntheis speech, server returned ${ res.status }`);\n }\n\n return res.arrayBuffer();\n}\n"],"mappings":";;;;;;;;;;;;;AAAA;;AAEA,IAAMA,gBAAgB,GAAG,OAAzB;AACA,IAAMC,aAAa,GAAG,gEAAtB;AACA,IAAMC,aAAa,GAAG,6CAAtB;;SAE8BC,e;;;;;6FAAf;IAAA;;IAAA;MAAA;QAAA;UAAA;YACbC,kBADa,QACbA,kBADa,mBAEbC,IAFa,EAEbA,IAFa,0BAENL,gBAFM,cAGbM,YAHa,QAGbA,YAHa,EAIbC,KAJa,QAIbA,KAJa,EAKbC,IALa,QAKbA,IALa,EAMbC,IANa,QAMbA,IANa,oBAObC,KAPa,EAObA,KAPa,2BAOLT,aAPK,eAQbU,MARa,QAQbA,MARa;YAUPC,IAVO,GAUA,IAAAC,kBAAA,EAAU;cAAER,IAAI,EAAJA,IAAF;cAAQE,KAAK,EAALA,KAAR;cAAeC,IAAI,EAAJA,IAAf;cAAqBC,IAAI,EAAJA,IAArB;cAA2BC,KAAK,EAALA,KAA3B;cAAkCC,MAAM,EAANA;YAAlC,CAAV,CAVA;YAAA;YAAA,OAYKG,KAAK,CAACZ,aAAD,EAAgB;cACrCa,OAAO,EAAE;gBACPC,aAAa,mBAAaZ,kBAAb,CADN;gBAEP,gBAAgB,sBAFT;gBAGP,4BAA4BE;cAHrB,CAD4B;cAMrCW,MAAM,EAAE,MAN6B;cAOrCC,IAAI,EAAEN;YAP+B,CAAhB,CAZV;;UAAA;YAYPO,GAZO;;YAAA,MAsBTA,GAAG,CAACC,MAAJ,KAAe,GAtBN;cAAA;cAAA;YAAA;;YAAA,MAuBL,IAAIC,KAAJ,sDAAyDF,GAAG,CAACC,MAA7D,EAvBK;;UAAA;YAAA,iCA0BND,GAAG,CAACG,WAAJ,EA1BM;;UAAA;UAAA;YAAA;QAAA;MAAA;IAAA;EAAA,C"}
|
|
@@ -1,335 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
|
|
3
|
-
var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
|
|
4
|
-
|
|
5
|
-
Object.defineProperty(exports, "__esModule", {
|
|
6
|
-
value: true
|
|
7
|
-
});
|
|
8
|
-
exports.default = _default;
|
|
9
|
-
|
|
10
|
-
var _SpeechSynthesisVoice = _interopRequireDefault(require("./SpeechSynthesisVoice"));
|
|
11
|
-
|
|
12
|
-
function _default() {
|
|
13
|
-
return [new _SpeechSynthesisVoice.default({
|
|
14
|
-
lang: 'ar-EG',
|
|
15
|
-
gender: 'Female',
|
|
16
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ar-EG, Hoda)'
|
|
17
|
-
}), new _SpeechSynthesisVoice.default({
|
|
18
|
-
lang: 'ar-SA',
|
|
19
|
-
gender: 'Male',
|
|
20
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ar-SA, Naayf)'
|
|
21
|
-
}), new _SpeechSynthesisVoice.default({
|
|
22
|
-
lang: 'bg-BG',
|
|
23
|
-
gender: 'Male',
|
|
24
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (bg-BG, Ivan)'
|
|
25
|
-
}), new _SpeechSynthesisVoice.default({
|
|
26
|
-
lang: 'ca-ES',
|
|
27
|
-
gender: 'Female',
|
|
28
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ca-ES, HerenaRUS)'
|
|
29
|
-
}), new _SpeechSynthesisVoice.default({
|
|
30
|
-
lang: 'cs-CZ',
|
|
31
|
-
gender: 'Male',
|
|
32
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (cs-CZ, Jakub)'
|
|
33
|
-
}), new _SpeechSynthesisVoice.default({
|
|
34
|
-
lang: 'da-DK',
|
|
35
|
-
gender: 'Female',
|
|
36
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (da-DK, HelleRUS)'
|
|
37
|
-
}), new _SpeechSynthesisVoice.default({
|
|
38
|
-
lang: 'de-AT',
|
|
39
|
-
gender: 'Male',
|
|
40
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (de-AT, Michael)'
|
|
41
|
-
}), new _SpeechSynthesisVoice.default({
|
|
42
|
-
lang: 'de-CH',
|
|
43
|
-
gender: 'Male',
|
|
44
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (de-CH, Karsten)'
|
|
45
|
-
}), new _SpeechSynthesisVoice.default({
|
|
46
|
-
lang: 'de-DE',
|
|
47
|
-
gender: 'Female',
|
|
48
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (de-DE, Hedda)'
|
|
49
|
-
}), new _SpeechSynthesisVoice.default({
|
|
50
|
-
lang: 'de-DE',
|
|
51
|
-
gender: 'Female',
|
|
52
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (de-DE, HeddaRUS)'
|
|
53
|
-
}), new _SpeechSynthesisVoice.default({
|
|
54
|
-
lang: 'de-DE',
|
|
55
|
-
gender: 'Male',
|
|
56
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (de-DE, Stefan, Apollo)'
|
|
57
|
-
}), new _SpeechSynthesisVoice.default({
|
|
58
|
-
lang: 'el-GR',
|
|
59
|
-
gender: 'Male',
|
|
60
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (el-GR, Stefanos)'
|
|
61
|
-
}), new _SpeechSynthesisVoice.default({
|
|
62
|
-
lang: 'en-AU',
|
|
63
|
-
gender: 'Female',
|
|
64
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-AU, Catherine)'
|
|
65
|
-
}), new _SpeechSynthesisVoice.default({
|
|
66
|
-
lang: 'en-AU',
|
|
67
|
-
gender: 'Female',
|
|
68
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-AU, HayleyRUS)'
|
|
69
|
-
}), new _SpeechSynthesisVoice.default({
|
|
70
|
-
lang: 'en-CA',
|
|
71
|
-
gender: 'Female',
|
|
72
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-CA, Linda)'
|
|
73
|
-
}), new _SpeechSynthesisVoice.default({
|
|
74
|
-
lang: 'en-CA',
|
|
75
|
-
gender: 'Female',
|
|
76
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-CA, HeatherRUS)'
|
|
77
|
-
}), new _SpeechSynthesisVoice.default({
|
|
78
|
-
lang: 'en-GB',
|
|
79
|
-
gender: 'Female',
|
|
80
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-GB, Susan, Apollo)'
|
|
81
|
-
}), new _SpeechSynthesisVoice.default({
|
|
82
|
-
lang: 'en-GB',
|
|
83
|
-
gender: 'Female',
|
|
84
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-GB, HazelRUS)'
|
|
85
|
-
}), new _SpeechSynthesisVoice.default({
|
|
86
|
-
lang: 'en-GB',
|
|
87
|
-
gender: 'Male',
|
|
88
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-GB, George, Apollo)'
|
|
89
|
-
}), new _SpeechSynthesisVoice.default({
|
|
90
|
-
lang: 'en-IE',
|
|
91
|
-
gender: 'Male',
|
|
92
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-IE, Sean)'
|
|
93
|
-
}), new _SpeechSynthesisVoice.default({
|
|
94
|
-
lang: 'en-IN',
|
|
95
|
-
gender: 'Female',
|
|
96
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-IN, Heera, Apollo)'
|
|
97
|
-
}), new _SpeechSynthesisVoice.default({
|
|
98
|
-
lang: 'en-IN',
|
|
99
|
-
gender: 'Female',
|
|
100
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-IN, PriyaRUS)'
|
|
101
|
-
}), new _SpeechSynthesisVoice.default({
|
|
102
|
-
lang: 'en-IN',
|
|
103
|
-
gender: 'Male',
|
|
104
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-IN, Ravi, Apollo)'
|
|
105
|
-
}), new _SpeechSynthesisVoice.default({
|
|
106
|
-
lang: 'en-US',
|
|
107
|
-
gender: 'Female',
|
|
108
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-US, ZiraRUS)'
|
|
109
|
-
}), new _SpeechSynthesisVoice.default({
|
|
110
|
-
lang: 'en-US',
|
|
111
|
-
gender: 'Female',
|
|
112
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-US, JessaRUS)'
|
|
113
|
-
}), new _SpeechSynthesisVoice.default({
|
|
114
|
-
lang: 'en-US',
|
|
115
|
-
gender: 'Male',
|
|
116
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-US, BenjaminRUS)'
|
|
117
|
-
}), new _SpeechSynthesisVoice.default({
|
|
118
|
-
lang: 'en-US',
|
|
119
|
-
gender: 'Female',
|
|
120
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-US, Jessa24kRUS)'
|
|
121
|
-
}), new _SpeechSynthesisVoice.default({
|
|
122
|
-
lang: 'en-US',
|
|
123
|
-
gender: 'Male',
|
|
124
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (en-US, Guy24kRUS)'
|
|
125
|
-
}), new _SpeechSynthesisVoice.default({
|
|
126
|
-
lang: 'es-ES',
|
|
127
|
-
gender: 'Female',
|
|
128
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (es-ES, Laura, Apollo)'
|
|
129
|
-
}), new _SpeechSynthesisVoice.default({
|
|
130
|
-
lang: 'es-ES',
|
|
131
|
-
gender: 'Female',
|
|
132
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (es-ES, HelenaRUS)'
|
|
133
|
-
}), new _SpeechSynthesisVoice.default({
|
|
134
|
-
lang: 'es-ES',
|
|
135
|
-
gender: 'Male',
|
|
136
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (es-ES, Pablo, Apollo)'
|
|
137
|
-
}), new _SpeechSynthesisVoice.default({
|
|
138
|
-
lang: 'es-MX',
|
|
139
|
-
gender: 'Female',
|
|
140
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (es-MX, HildaRUS)'
|
|
141
|
-
}), new _SpeechSynthesisVoice.default({
|
|
142
|
-
lang: 'es-MX',
|
|
143
|
-
gender: 'Male',
|
|
144
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (es-MX, Raul, Apollo)'
|
|
145
|
-
}), new _SpeechSynthesisVoice.default({
|
|
146
|
-
lang: 'fi-FI',
|
|
147
|
-
gender: 'Female',
|
|
148
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (fi-FI, HeidiRUS)'
|
|
149
|
-
}), new _SpeechSynthesisVoice.default({
|
|
150
|
-
lang: 'fr-CA',
|
|
151
|
-
gender: 'Female',
|
|
152
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (fr-CA, Caroline)'
|
|
153
|
-
}), new _SpeechSynthesisVoice.default({
|
|
154
|
-
lang: 'fr-CA',
|
|
155
|
-
gender: 'Female',
|
|
156
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (fr-CA, HarmonieRUS)'
|
|
157
|
-
}), new _SpeechSynthesisVoice.default({
|
|
158
|
-
lang: 'fr-CH',
|
|
159
|
-
gender: 'Male',
|
|
160
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (fr-CH, Guillaume)'
|
|
161
|
-
}), new _SpeechSynthesisVoice.default({
|
|
162
|
-
lang: 'fr-FR',
|
|
163
|
-
gender: 'Female',
|
|
164
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (fr-FR, Julie, Apollo)'
|
|
165
|
-
}), new _SpeechSynthesisVoice.default({
|
|
166
|
-
lang: 'fr-FR',
|
|
167
|
-
gender: 'Female',
|
|
168
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (fr-FR, HortenseRUS)'
|
|
169
|
-
}), new _SpeechSynthesisVoice.default({
|
|
170
|
-
lang: 'fr-FR',
|
|
171
|
-
gender: 'Male',
|
|
172
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (fr-FR, Paul, Apollo)'
|
|
173
|
-
}), new _SpeechSynthesisVoice.default({
|
|
174
|
-
lang: 'he-IL',
|
|
175
|
-
gender: 'Male',
|
|
176
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (he-IL, Asaf)'
|
|
177
|
-
}), new _SpeechSynthesisVoice.default({
|
|
178
|
-
lang: 'hi-IN',
|
|
179
|
-
gender: 'Female',
|
|
180
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (hi-IN, Kalpana, Apollo)'
|
|
181
|
-
}), new _SpeechSynthesisVoice.default({
|
|
182
|
-
lang: 'hi-IN',
|
|
183
|
-
gender: 'Female',
|
|
184
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (hi-IN, Kalpana)'
|
|
185
|
-
}), new _SpeechSynthesisVoice.default({
|
|
186
|
-
lang: 'hi-IN',
|
|
187
|
-
gender: 'Male',
|
|
188
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (hi-IN, Hemant)'
|
|
189
|
-
}), new _SpeechSynthesisVoice.default({
|
|
190
|
-
lang: 'hr-HR',
|
|
191
|
-
gender: 'Male',
|
|
192
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (hr-HR, Matej)'
|
|
193
|
-
}), new _SpeechSynthesisVoice.default({
|
|
194
|
-
lang: 'hu-HU',
|
|
195
|
-
gender: 'Male',
|
|
196
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (hu-HU, Szabolcs)'
|
|
197
|
-
}), new _SpeechSynthesisVoice.default({
|
|
198
|
-
lang: 'id-ID',
|
|
199
|
-
gender: 'Male',
|
|
200
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (id-ID, Andika)'
|
|
201
|
-
}), new _SpeechSynthesisVoice.default({
|
|
202
|
-
lang: 'it-IT',
|
|
203
|
-
gender: 'Male',
|
|
204
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (it-IT, Cosimo, Apollo)'
|
|
205
|
-
}), new _SpeechSynthesisVoice.default({
|
|
206
|
-
lang: 'it-IT',
|
|
207
|
-
gender: 'Female',
|
|
208
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (it-IT, LuciaRUS)'
|
|
209
|
-
}), new _SpeechSynthesisVoice.default({
|
|
210
|
-
lang: 'ja-JP',
|
|
211
|
-
gender: 'Female',
|
|
212
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ja-JP, Ayumi, Apollo)'
|
|
213
|
-
}), new _SpeechSynthesisVoice.default({
|
|
214
|
-
lang: 'ja-JP',
|
|
215
|
-
gender: 'Male',
|
|
216
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ja-JP, Ichiro, Apollo)'
|
|
217
|
-
}), new _SpeechSynthesisVoice.default({
|
|
218
|
-
lang: 'ja-JP',
|
|
219
|
-
gender: 'Female',
|
|
220
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ja-JP, HarukaRUS)'
|
|
221
|
-
}), new _SpeechSynthesisVoice.default({
|
|
222
|
-
lang: 'ko-KR',
|
|
223
|
-
gender: 'Female',
|
|
224
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ko-KR, HeamiRUS)'
|
|
225
|
-
}), new _SpeechSynthesisVoice.default({
|
|
226
|
-
lang: 'ms-MY',
|
|
227
|
-
gender: 'Male',
|
|
228
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ms-MY, Rizwan)'
|
|
229
|
-
}), new _SpeechSynthesisVoice.default({
|
|
230
|
-
lang: 'nb-NO',
|
|
231
|
-
gender: 'Female',
|
|
232
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (nb-NO, HuldaRUS)'
|
|
233
|
-
}), new _SpeechSynthesisVoice.default({
|
|
234
|
-
lang: 'nl-NL',
|
|
235
|
-
gender: 'Female',
|
|
236
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (nl-NL, HannaRUS)'
|
|
237
|
-
}), new _SpeechSynthesisVoice.default({
|
|
238
|
-
lang: 'pl-PL',
|
|
239
|
-
gender: 'Female',
|
|
240
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (pl-PL, PaulinaRUS)'
|
|
241
|
-
}), new _SpeechSynthesisVoice.default({
|
|
242
|
-
lang: 'pt-BR',
|
|
243
|
-
gender: 'Female',
|
|
244
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (pt-BR, HeloisaRUS)'
|
|
245
|
-
}), new _SpeechSynthesisVoice.default({
|
|
246
|
-
lang: 'pt-BR',
|
|
247
|
-
gender: 'Male',
|
|
248
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (pt-BR, Daniel, Apollo)'
|
|
249
|
-
}), new _SpeechSynthesisVoice.default({
|
|
250
|
-
lang: 'pt-PT',
|
|
251
|
-
gender: 'Female',
|
|
252
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (pt-PT, HeliaRUS)'
|
|
253
|
-
}), new _SpeechSynthesisVoice.default({
|
|
254
|
-
lang: 'ro-RO',
|
|
255
|
-
gender: 'Male',
|
|
256
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ro-RO, Andrei)'
|
|
257
|
-
}), new _SpeechSynthesisVoice.default({
|
|
258
|
-
lang: 'ru-RU',
|
|
259
|
-
gender: 'Female',
|
|
260
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ru-RU, Irina, Apollo)'
|
|
261
|
-
}), new _SpeechSynthesisVoice.default({
|
|
262
|
-
lang: 'ru-RU',
|
|
263
|
-
gender: 'Male',
|
|
264
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ru-RU, Pavel, Apollo)'
|
|
265
|
-
}), new _SpeechSynthesisVoice.default({
|
|
266
|
-
lang: 'ru-RU',
|
|
267
|
-
gender: 'Female',
|
|
268
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ru-RU, EkaterinaRUS)'
|
|
269
|
-
}), new _SpeechSynthesisVoice.default({
|
|
270
|
-
lang: 'sk-SK',
|
|
271
|
-
gender: 'Male',
|
|
272
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (sk-SK, Filip)'
|
|
273
|
-
}), new _SpeechSynthesisVoice.default({
|
|
274
|
-
lang: 'sl-SI',
|
|
275
|
-
gender: 'Male',
|
|
276
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (sl-SI, Lado)'
|
|
277
|
-
}), new _SpeechSynthesisVoice.default({
|
|
278
|
-
lang: 'sv-SE',
|
|
279
|
-
gender: 'Female',
|
|
280
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (sv-SE, HedvigRUS)'
|
|
281
|
-
}), new _SpeechSynthesisVoice.default({
|
|
282
|
-
lang: 'ta-IN',
|
|
283
|
-
gender: 'Male',
|
|
284
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (ta-IN, Valluvar)'
|
|
285
|
-
}), new _SpeechSynthesisVoice.default({
|
|
286
|
-
lang: 'th-TH',
|
|
287
|
-
gender: 'Male',
|
|
288
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (th-TH, Pattara)'
|
|
289
|
-
}), new _SpeechSynthesisVoice.default({
|
|
290
|
-
lang: 'tr-TR',
|
|
291
|
-
gender: 'Female',
|
|
292
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (tr-TR, SedaRUS)'
|
|
293
|
-
}), new _SpeechSynthesisVoice.default({
|
|
294
|
-
lang: 'vi-VN',
|
|
295
|
-
gender: 'Male',
|
|
296
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (vi-VN, An)'
|
|
297
|
-
}), new _SpeechSynthesisVoice.default({
|
|
298
|
-
lang: 'zh-CN',
|
|
299
|
-
gender: 'Female',
|
|
300
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-CN, HuihuiRUS)'
|
|
301
|
-
}), new _SpeechSynthesisVoice.default({
|
|
302
|
-
lang: 'zh-CN',
|
|
303
|
-
gender: 'Female',
|
|
304
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-CN, Yaoyao, Apollo)'
|
|
305
|
-
}), new _SpeechSynthesisVoice.default({
|
|
306
|
-
lang: 'zh-CN',
|
|
307
|
-
gender: 'Male',
|
|
308
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-CN, Kangkang, Apollo)'
|
|
309
|
-
}), new _SpeechSynthesisVoice.default({
|
|
310
|
-
lang: 'zh-HK',
|
|
311
|
-
gender: 'Female',
|
|
312
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-HK, Tracy, Apollo)'
|
|
313
|
-
}), new _SpeechSynthesisVoice.default({
|
|
314
|
-
lang: 'zh-HK',
|
|
315
|
-
gender: 'Female',
|
|
316
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-HK, TracyRUS)'
|
|
317
|
-
}), new _SpeechSynthesisVoice.default({
|
|
318
|
-
lang: 'zh-HK',
|
|
319
|
-
gender: 'Male',
|
|
320
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-HK, Danny, Apollo)'
|
|
321
|
-
}), new _SpeechSynthesisVoice.default({
|
|
322
|
-
lang: 'zh-TW',
|
|
323
|
-
gender: 'Female',
|
|
324
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-TW, Yating, Apollo)'
|
|
325
|
-
}), new _SpeechSynthesisVoice.default({
|
|
326
|
-
lang: 'zh-TW',
|
|
327
|
-
gender: 'Female',
|
|
328
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-TW, HanHanRUS)'
|
|
329
|
-
}), new _SpeechSynthesisVoice.default({
|
|
330
|
-
lang: 'zh-TW',
|
|
331
|
-
gender: 'Male',
|
|
332
|
-
voiceURI: 'Microsoft Server Speech Text to Speech Voice (zh-TW, Zhiwei, Apollo)'
|
|
333
|
-
})];
|
|
334
|
-
}
|
|
335
|
-
//# sourceMappingURL=fetchVoices.js.map
|