const Model = require('objection').Model const path = require('path') const fs = require('fs-extra') const _ = require('lodash') const yaml = require('js-yaml') const commonHelper = require('../helpers/common') /* global WIKI */ /** * SearchEngine model */ module.exports = class SearchEngine extends Model { static get tableName() { return 'searchEngines' } static get idColumn() { return 'key' } static get jsonSchema () { return { type: 'object', required: ['key', 'isEnabled'], properties: { key: {type: 'string'}, isEnabled: {type: 'boolean'}, level: {type: 'string'} } } } static get jsonAttributes() { return ['config'] } static async getSearchEngines() { return WIKI.models.searchEngines.query() } static async refreshSearchEnginesFromDisk() { let trx try { const dbSearchEngines = await WIKI.models.searchEngines.query() // -> Fetch definitions from disk const searchEnginesDirs = await fs.readdir(path.join(WIKI.SERVERPATH, 'modules/search')) let diskSearchEngines = [] for (let dir of searchEnginesDirs) { const def = await fs.readFile(path.join(WIKI.SERVERPATH, 'modules/search', dir, 'definition.yml'), 'utf8') diskSearchEngines.push(yaml.safeLoad(def)) } WIKI.data.searchEngines = diskSearchEngines.map(searchEngine => ({ ...searchEngine, props: commonHelper.parseModuleProps(searchEngine.props) })) // -> Insert new searchEngines let newSearchEngines = [] for (let searchEngine of WIKI.data.searchEngines) { if (!_.some(dbSearchEngines, ['key', searchEngine.key])) { newSearchEngines.push({ key: searchEngine.key, isEnabled: false, config: _.transform(searchEngine.props, (result, value, key) => { _.set(result, key, value.default) return result }, {}) }) } else { const searchEngineConfig = _.get(_.find(dbSearchEngines, ['key', searchEngine.key]), 'config', {}) await WIKI.models.searchEngines.query().patch({ config: _.transform(searchEngine.props, (result, value, key) => { if (!_.has(result, key)) { _.set(result, key, value.default) } return result }, searchEngineConfig) }).where('key', searchEngine.key) } } if (newSearchEngines.length > 0) { trx = await WIKI.models.Objection.transaction.start(WIKI.models.knex) for (let searchEngine of newSearchEngines) { await WIKI.models.searchEngines.query(trx).insert(searchEngine) } await trx.commit() WIKI.logger.info(`Loaded ${newSearchEngines.length} new search engines: [ OK ]`) } else { WIKI.logger.info(`No new search engines found: [ SKIPPED ]`) } } catch (err) { WIKI.logger.error(`Failed to scan or load new search engines: [ FAILED ]`) WIKI.logger.error(err) if (trx) { trx.rollback() } } } static async initEngine() { const searchEngine = await WIKI.models.searchEngines.query().findOne('isEnabled', true) if (searchEngine) { WIKI.data.searchEngine = require(`../modules/search/${searchEngine.key}/engine`) WIKI.data.searchEngine.config = searchEngine.config try { await WIKI.data.searchEngine.init() } catch (err) { WIKI.logger.warn(err) } } } static async pageEvent({ event, page }) { const searchEngines = await WIKI.models.storage.query().where('isEnabled', true) if (searchEngines && searchEngines.length > 0) { _.forEach(searchEngines, logger => { WIKI.queue.job.syncStorage.add({ event, logger, page }, { removeOnComplete: true }) }) } } }