yomichan/ext/mixed/js/audio-system.js

138 lines
4.3 KiB
JavaScript
Raw Normal View History

2019-10-10 23:58:06 +00:00
/*
* Copyright (C) 2019-2021 Yomichan Authors
2019-10-10 23:58:06 +00:00
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
2020-01-01 17:00:31 +00:00
* along with this program. If not, see <https://www.gnu.org/licenses/>.
2019-10-10 23:58:06 +00:00
*/
/* global
* CacheMap
* TextToSpeechAudio
* api
*/
2020-03-07 17:44:14 +00:00
class AudioSystem {
constructor(useCache) {
this._cache = new CacheMap(useCache ? 32 : 0);
this._fallbackAudio = null;
}
2020-03-07 17:44:14 +00:00
prepare() {
// speechSynthesis.getVoices() will not be populated unless some API call is made.
if (typeof speechSynthesis === 'undefined') { return; }
const eventListeners = new EventListenerCollection();
const onVoicesChanged = () => { eventListeners.removeAllEventListeners(); };
eventListeners.addEventListener(speechSynthesis, 'voiceschanged', onVoicesChanged, false);
2020-03-07 17:44:14 +00:00
}
async createExpressionAudio(sources, expression, reading, details) {
const key = JSON.stringify([expression, reading]);
const cacheValue = this._cache.get(key);
if (typeof cacheValue !== 'undefined') {
return cacheValue;
2020-03-07 17:44:14 +00:00
}
for (let i = 0, ii = sources.length; i < ii; ++i) {
const source = sources[i];
const infoList = await await api.getExpressionAudioInfoList(source, expression, reading, details);
for (let j = 0, jj = infoList.length; j < jj; ++j) {
const info = infoList[j];
let audio;
try {
audio = await this.createAudioFromInfo(info, source);
} catch (e) {
continue;
}
2020-03-07 17:44:14 +00:00
const result = {audio, source, infoList, infoListIndex: j};
this._cache.set(key, result);
return result;
}
}
throw new Error('Could not create audio');
}
getFallbackAudio() {
if (this._fallbackAudio === null) {
this._fallbackAudio = new Audio('/mixed/mp3/button.mp3');
}
return this._fallbackAudio;
}
createAudio(url, source) {
2020-03-07 17:44:14 +00:00
return new Promise((resolve, reject) => {
const audio = new Audio(url);
audio.addEventListener('loadeddata', () => {
if (!this._isAudioValid(audio, source)) {
2020-03-07 17:44:14 +00:00
reject(new Error('Could not retrieve audio'));
} else {
resolve(audio);
}
});
audio.addEventListener('error', () => reject(audio.error));
});
}
createTextToSpeechAudio(text, voiceUri) {
const voice = this._getTextToSpeechVoiceFromVoiceUri(voiceUri);
if (voice === null) {
throw new Error('Invalid text-to-speech voice');
}
return new TextToSpeechAudio(text, voice);
}
async createAudioFromInfo(info, source) {
switch (info.type) {
case 'url':
return await this.createAudio(info.url, source);
case 'tts':
return this.createTextToSpeechAudio(info.text, info.voice);
default:
throw new Error(`Unsupported type: ${info.type}`);
}
}
// Private
_isAudioValid(audio, source) {
switch (source) {
case 'jpod101':
{
const duration = audio.duration;
return (
duration !== 5.694694 && // Invalid audio (Chrome)
duration !== 5.720718 // Invalid audio (Firefox)
);
}
default:
return true;
}
}
2020-03-07 17:44:14 +00:00
_getTextToSpeechVoiceFromVoiceUri(voiceUri) {
try {
for (const voice of speechSynthesis.getVoices()) {
if (voice.voiceURI === voiceUri) {
return voice;
}
}
} catch (e) {
// NOP
}
return null;
}
}