|
|
|
const Model = require('objection').Model
|
|
|
|
const path = require('path')
|
|
|
|
const fs = require('fs-extra')
|
|
|
|
const _ = require('lodash')
|
|
|
|
const yaml = require('js-yaml')
|
|
|
|
const commonHelper = require('../helpers/common')
|
|
|
|
|
|
|
|
/* global WIKI */
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Logger model
|
|
|
|
*/
|
|
|
|
module.exports = class Logger extends Model {
|
|
|
|
static get tableName() { return 'loggers' }
|
|
|
|
static get idColumn() { return 'key' }
|
|
|
|
|
|
|
|
static get jsonSchema () {
|
|
|
|
return {
|
|
|
|
type: 'object',
|
|
|
|
required: ['key', 'isEnabled'],
|
|
|
|
|
|
|
|
properties: {
|
|
|
|
key: {type: 'string'},
|
|
|
|
isEnabled: {type: 'boolean'},
|
|
|
|
level: {type: 'string'}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static get jsonAttributes() {
|
|
|
|
return ['config']
|
|
|
|
}
|
|
|
|
|
|
|
|
static async getLoggers() {
|
|
|
|
return WIKI.models.loggers.query()
|
|
|
|
}
|
|
|
|
|
|
|
|
static async refreshLoggersFromDisk() {
|
|
|
|
let trx
|
|
|
|
try {
|
|
|
|
const dbLoggers = await WIKI.models.loggers.query()
|
|
|
|
|
|
|
|
// -> Fetch definitions from disk
|
|
|
|
const loggersDirs = await fs.readdir(path.join(WIKI.SERVERPATH, 'modules/logging'))
|
|
|
|
let diskLoggers = []
|
|
|
|
for (let dir of loggersDirs) {
|
|
|
|
const def = await fs.readFile(path.join(WIKI.SERVERPATH, 'modules/logging', dir, 'definition.yml'), 'utf8')
|
|
|
|
diskLoggers.push(yaml.safeLoad(def))
|
|
|
|
}
|
|
|
|
WIKI.data.loggers = diskLoggers.map(logger => ({
|
|
|
|
...logger,
|
|
|
|
props: commonHelper.parseModuleProps(logger.props)
|
|
|
|
}))
|
|
|
|
|
|
|
|
// -> Insert new loggers
|
|
|
|
let newLoggers = []
|
|
|
|
for (let logger of WIKI.data.loggers) {
|
|
|
|
if (!_.some(dbLoggers, ['key', logger.key])) {
|
|
|
|
newLoggers.push({
|
|
|
|
key: logger.key,
|
|
|
|
isEnabled: (logger.key === 'console'),
|
|
|
|
level: logger.defaultLevel,
|
|
|
|
config: _.transform(logger.props, (result, value, key) => {
|
|
|
|
_.set(result, key, value.default)
|
|
|
|
return result
|
|
|
|
}, {})
|
|
|
|
})
|
|
|
|
} else {
|
|
|
|
const loggerConfig = _.get(_.find(dbLoggers, ['key', logger.key]), 'config', {})
|
|
|
|
await WIKI.models.loggers.query().patch({
|
|
|
|
config: _.transform(logger.props, (result, value, key) => {
|
|
|
|
if (!_.has(result, key)) {
|
|
|
|
_.set(result, key, value.default)
|
|
|
|
}
|
|
|
|
return result
|
|
|
|
}, loggerConfig)
|
|
|
|
}).where('key', logger.key)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (newLoggers.length > 0) {
|
|
|
|
trx = await WIKI.models.Objection.transaction.start(WIKI.models.knex)
|
|
|
|
for (let logger of newLoggers) {
|
|
|
|
await WIKI.models.loggers.query(trx).insert(logger)
|
|
|
|
}
|
|
|
|
await trx.commit()
|
|
|
|
WIKI.logger.info(`Loaded ${newLoggers.length} new loggers: [ OK ]`)
|
|
|
|
} else {
|
|
|
|
WIKI.logger.info(`No new loggers found: [ SKIPPED ]`)
|
|
|
|
}
|
|
|
|
} catch (err) {
|
|
|
|
WIKI.logger.error(`Failed to scan or load new loggers: [ FAILED ]`)
|
|
|
|
WIKI.logger.error(err)
|
|
|
|
if (trx) {
|
|
|
|
trx.rollback()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static async pageEvent({ event, page }) {
|
|
|
|
const loggers = await WIKI.models.storage.query().where('isEnabled', true)
|
|
|
|
if (loggers && loggers.length > 0) {
|
|
|
|
_.forEach(loggers, logger => {
|
|
|
|
WIKI.queue.job.syncStorage.add({
|
|
|
|
event,
|
|
|
|
logger,
|
|
|
|
page
|
|
|
|
}, {
|
|
|
|
removeOnComplete: true
|
|
|
|
})
|
|
|
|
})
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|