/** * AssetGuard * * This module aims to provide a comprehensive and stable method for processing * and downloading game assets for the WesterosCraft server. A central object * stores download meta for several identifiers (categories). This meta data * is initially empty until one of the module's processing functions are called. * That function will process the corresponding asset index and validate any exisitng * local files. If a file is missing or fails validation, it will be placed into an * array which acts as a queue. This queue is wrapped in a download tracker object * so that essential information can be cached. The download tracker object is then * assigned as the value of the identifier in the central object. These download * trackers will remain idle until an async process is started to process them. * * Once the async process is started, any enqueued assets will be downloaded. The central * object will emit events throughout the download whose name correspond to the identifier * being processed. For example, if the 'assets' identifier was being processed, whenever * the download stream recieves data, the event 'assetsdlprogress' will be emitted off of * the central object instance. This can be listened to by external modules allowing for * categorical tracking of the downloading process. * * @module assetguard */ // Requirements const fs = require('fs') const request = require('request') const path = require('path') const mkpath = require('mkdirp'); const async = require('async') const crypto = require('crypto') const AdmZip = require('adm-zip') const child_process = require('child_process') const EventEmitter = require('events') const {remote} = require('electron') // Classes /** Class representing a base asset. */ class Asset{ /** * Create an asset. * * @param {any} id - id of the asset. * @param {String} hash - hash value of the asset. * @param {Number} size - size in bytes of the asset. * @param {String} from - url where the asset can be found. * @param {String} to - absolute local file path of the asset. */ constructor(id, hash, size, from, to){ this.id = id this.hash = hash this.size = size this.from = from this.to = to } } /** Class representing a mojang library. */ class Library extends Asset{ /** * Converts the process.platform OS names to match mojang's OS names. */ static mojangFriendlyOS(){ const opSys = process.platform if (opSys === 'darwin') { return 'osx'; } else if (opSys === 'win32'){ return 'windows'; } else if (opSys === 'linux'){ return 'linux'; } else { return 'unknown_os'; } } /** * Checks whether or not a library is valid for download on a particular OS, following * the rule format specified in the mojang version data index. If the allow property has * an OS specified, then the library can ONLY be downloaded on that OS. If the disallow * property has instead specified an OS, the library can be downloaded on any OS EXCLUDING * the one specified. * * @param {Object} rules - the Library's download rules. * @returns {Boolean} - true if the Library follows the specified rules, otherwise false. */ static validateRules(rules){ if(rules == null) return true let result = true rules.forEach(function(rule){ const action = rule['action'] const osProp = rule['os'] if(action != null){ if(osProp != null){ const osName = osProp['name'] const osMoj = Library.mojangFriendlyOS() if(action === 'allow'){ result = osName === osMoj return } else if(action === 'disallow'){ result = osName !== osMoj return } } } }) return result } } class DistroModule extends Asset { /** * Create a DistroModule. This is for processing, * not equivalent to the module objects in the * distro index. * * @param {any} id - id of the asset. * @param {String} hash - hash value of the asset. * @param {Number} size - size in bytes of the asset. * @param {String} from - url where the asset can be found. * @param {String} to - absolute local file path of the asset. * @param {String} type - the module type. */ constructor(id, hash, size, from, to, type){ super(id, hash, size, from, to) this.type = type } } /** * Class representing a download tracker. This is used to store meta data * about a download queue, including the queue itself. */ class DLTracker { /** * Create a DLTracker * * @param {Array.} dlqueue - an array containing assets queued for download. * @param {Number} dlsize - the combined size of each asset in the download queue array. * @param {function(Asset)} callback - optional callback which is called when an asset finishes downloading. */ constructor(dlqueue, dlsize, callback = null){ this.dlqueue = dlqueue this.dlsize = dlsize this.callback = callback } } /** * Central object class used for control flow. This object stores data about * categories of downloads. Each category is assigned an identifier with a * DLTracker object as its value. Combined information is also stored, such as * the total size of all the queued files in each category. This event is used * to emit events so that external modules can listen into processing done in * this module. */ class AssetGuard extends EventEmitter{ /** * AssetGuard class should only ever have one instance which is defined in * this module. On creation the object's properties are never-null default * values. Each identifier is resolved to an empty DLTracker. */ constructor(){ super() this.totaldlsize = 0; this.progress = 0; this.assets = new DLTracker([], 0) this.libraries = new DLTracker([], 0) this.files = new DLTracker([], 0) this.forge = new DLTracker([], 0) } } /** * Global static final instance of AssetGuard */ const instance = new AssetGuard() // Utility Functions /** * Resolve an artifact id into a path. For example, on windows * 'net.minecraftforge:forge:1.11.2-13.20.0.2282', '.jar' becomes * net\minecraftforge\forge\1.11.2-13.20.0.2282\forge-1.11.2-13.20.0.2282.jar * * @param {String} artifactid - the artifact id string. * @param {String} extension - the extension of the file at the resolved path. * @returns {String} - the resolved relative path from the artifact id. */ function _resolvePath(artifactid, extension){ let ps = artifactid.split(':') let cs = ps[0].split('.') cs.push(ps[1]) cs.push(ps[2]) cs.push(ps[1].concat('-').concat(ps[2]).concat(extension)) return path.join.apply(path, cs) } /** * Resolve an artifact id into a URL. For example, * 'net.minecraftforge:forge:1.11.2-13.20.0.2282', '.jar' becomes * net/minecraftforge/forge/1.11.2-13.20.0.2282/forge-1.11.2-13.20.0.2282.jar * * @param {String} artifactid - the artifact id string. * @param {String} extension - the extension of the file at the resolved url. * @returns {String} - the resolved relative URL from the artifact id. */ function _resolveURL(artifactid, extension){ let ps = artifactid.split(':') let cs = ps[0].split('.') cs.push(ps[1]) cs.push(ps[2]) cs.push(ps[1].concat('-').concat(ps[2]).concat(extension)) return cs.join('/') } /** * Calculates the hash for a file using the specified algorithm. * * @param {Buffer} buf - the buffer containing file data. * @param {String} algo - the hash algorithm. * @returns {String} - the calculated hash in hex. */ function _calculateHash(buf, algo){ return crypto.createHash(algo).update(buf).digest('hex') } /** * Used to parse a checksums file. This is specifically designed for * the checksums.sha1 files found inside the forge scala dependencies. * * @param {String} content - the string content of the checksums file. * @returns {Object} - an object with keys being the file names, and values being the hashes. */ function _parseChecksumsFile(content){ let finalContent = {} let lines = content.split('\n') for(let i=0; i} checksums - the checksums listed in the forge version index. * @returns {Boolean} - true if the file exists and the hashes match, otherwise false. */ function _validateForgeChecksum(filePath, checksums){ if(fs.existsSync(filePath)){ if(checksums == null || checksums.length === 0){ return true } let buf = fs.readFileSync(filePath) let calcdhash = _calculateHash(buf, 'sha1') let valid = checksums.includes(calcdhash) if(!valid && filePath.endsWith('.jar')){ valid = _validateForgeJar(filePath, checksums) } return valid } return false } /** * Validates a forge jar file dependency who declares a checksums.sha1 file. * This can be an expensive task as it usually requires that we calculate thousands * of hashes. * * @param {Buffer} buf - the buffer of the jar file. * @param {Array.} checksums - the checksums listed in the forge version index. * @returns {Boolean} - true if all hashes declared in the checksums.sha1 file match the actual hashes. */ function _validateForgeJar(buf, checksums){ // Double pass method was the quickest I found. I tried a version where we store data // to only require a single pass, plus some quick cleanup but that seemed to take slightly more time. const hashes = {} let expected = {} const zip = new AdmZip(buf) const zipEntries = zip.getEntries() //First pass for(let i=0; i} filePaths - The paths of the files to be extracted and unpacked. * @returns {Promise.} - An empty promise to indicate the extraction has completed. */ function _extractPackXZ(filePaths){ return new Promise(function(fulfill, reject){ const libPath = path.join(__dirname, '..', 'libraries', 'java', 'PackXZExtract.jar') const filePath = filePaths.join(',') const child = child_process.spawn('C:\\Program Files\\Java\\jre1.8.0_131\\bin\\javaw.exe', ['-jar', libPath, '-packxz', filePath]) child.stdout.on('data', (data) => { //console.log('PackXZExtract:', data.toString('utf8')) }) child.stderr.on('data', (data) => { //console.log('PackXZExtract:', data.toString('utf8')) }) child.on('close', (code, signal) => { //console.log('PackXZExtract: Exited with code', code) fulfill() }) }) } /** * Function which finalizes the forge installation process. This creates a 'version' * instance for forge and saves its version.json file into that instance. If that * instance already exists, the contents of the version.json file are read and returned * in a promise. * * @param {Asset} asset - The Asset object representing Forge. * @param {String} basePath * @returns {Promise.} - A promise which resolves to the contents of forge's version.json. */ function _finalizeForgeAsset(asset, basePath){ return new Promise(function(fulfill, reject){ fs.readFile(asset.to, (err, data) => { const zip = new AdmZip(data) const zipEntries = zip.getEntries() for(let i=0; i { if(resp.statusCode === 200){ let writeStream = fs.createWriteStream(asset.to) writeStream.on('close', () => { //console.log('DLResults ' + asset.size + ' ' + count + ' ', asset.size === count) if(concurrentDlTracker.callback != null){ concurrentDlTracker.callback.apply(concurrentDlTracker, [asset]) } cb() }) req.pipe(writeStream) req.resume() } else { req.abort() console.log('Failed to download ' + asset.from + '. Response code', resp.statusCode) instance.progress += asset.size*1 win.setProgressBar(instance.progress/instance.totaldlsize) cb() } }) req.on('data', function(chunk){ count += chunk.length instance.progress += chunk.length acc += chunk.length instance.emit(identifier + 'dlprogress', acc) //console.log(identifier + ' Progress', acc/instance[identifier].dlsize) win.setProgressBar(instance.progress/instance.totaldlsize) }) }, function(err){ if(err){ instance.emit(identifier + 'dlerror') console.log('An item in ' + identifier + ' failed to process'); } else { instance.emit(identifier + 'dlcomplete') console.log('All ' + identifier + ' have been processed successfully') } instance.totaldlsize -= instance[identifier].dlsize instance.progress -= instance[identifier].dlsize instance[identifier] = new DLTracker([], 0) if(instance.totaldlsize === 0) { win.setProgressBar(-1) instance.emit('dlcomplete') } }) return true } } // Validation Functions /** * Loads the version data for a given minecraft version. * * @param {String} version - the game version for which to load the index data. * @param {String} basePath - the absolute file path which will be prepended to the given relative paths. * @param {Boolean} force - optional. If true, the version index will be downloaded even if it exists locally. Defaults to false. * @returns {Promise.} - Promise which resolves to the version data object. */ function loadVersionData(version, basePath, force = false){ return new Promise(function(fulfill, reject){ const name = version + '.json' const url = 'https://s3.amazonaws.com/Minecraft.Download/versions/' + version + '/' + name const versionPath = path.join(basePath, 'versions', version) const versionFile = path.join(versionPath, name) if(!fs.existsSync(versionFile) || force){ //This download will never be tracked as it's essential and trivial. request.head(url, function(err, res, body){ console.log('Preparing download of ' + version + ' assets.') mkpath.sync(versionPath) const stream = request(url).pipe(fs.createWriteStream(versionFile)) stream.on('finish', function(){ fulfill(JSON.parse(fs.readFileSync(versionFile))) }) }) } else { fulfill(JSON.parse(fs.readFileSync(versionFile))) } }) } /** * Public asset validation function. This function will handle the validation of assets. * It will parse the asset index specified in the version data, analyzing each * asset entry. In this analysis it will check to see if the local file exists and is valid. * If not, it will be added to the download queue for the 'assets' identifier. * * @param {Object} versionData - the version data for the assets. * @param {String} basePath - the absolute file path which will be prepended to the given relative paths. * @param {Boolean} force - optional. If true, the asset index will be downloaded even if it exists locally. Defaults to false. * @returns {Promise.} - An empty promise to indicate the async processing has completed. */ function validateAssets(versionData, basePath, force = false){ return new Promise(function(fulfill, reject){ _assetChainIndexData(versionData, basePath, force).then(() => { fulfill() }) }) } //Chain the asset tasks to provide full async. The below functions are private. /** * Private function used to chain the asset validation process. This function retrieves * the index data. * @param {Object} versionData * @param {String} basePath * @param {Boolean} force * @returns {Promise.} - An empty promise to indicate the async processing has completed. */ function _assetChainIndexData(versionData, basePath, force = false){ return new Promise(function(fulfill, reject){ //Asset index constants. const assetIndex = versionData.assetIndex const name = assetIndex.id + '.json' const indexPath = path.join(basePath, 'assets', 'indexes') const assetIndexLoc = path.join(indexPath, name) let data = null if(!fs.existsSync(assetIndexLoc) || force){ console.log('Downloading ' + versionData.id + ' asset index.') mkpath.sync(indexPath) const stream = request(assetIndex.url).pipe(fs.createWriteStream(assetIndexLoc)) stream.on('finish', function() { data = JSON.parse(fs.readFileSync(assetIndexLoc, 'utf-8')) _assetChainValidateAssets(versionData, basePath, data).then(() => { fulfill() }) }) } else { data = JSON.parse(fs.readFileSync(assetIndexLoc, 'utf-8')) _assetChainValidateAssets(versionData, basePath, data).then(() => { fulfill() }) } }) } /** * Private function used to chain the asset validation process. This function processes * the assets and enqueues missing or invalid files. * @param {Object} versionData * @param {String} basePath * @param {Boolean} force * @returns {Promise.} - An empty promise to indicate the async processing has completed. */ function _assetChainValidateAssets(versionData, basePath, indexData){ return new Promise(function(fulfill, reject){ //Asset constants const resourceURL = 'http://resources.download.minecraft.net/' const localPath = path.join(basePath, 'assets') const indexPath = path.join(localPath, 'indexes') const objectPath = path.join(localPath, 'objects') const assetDlQueue = [] let dlSize = 0; //const objKeys = Object.keys(data.objects) async.forEachOfLimit(indexData.objects, 10, function(value, key, cb){ const hash = value.hash const assetName = path.join(hash.substring(0, 2), hash) const urlName = hash.substring(0, 2) + "/" + hash const ast = new Asset(key, hash, String(value.size), resourceURL + urlName, path.join(objectPath, assetName)) if(!_validateLocal(ast.to, 'sha1', ast.hash)){ dlSize += (ast.size*1) assetDlQueue.push(ast) } cb() }, function(err){ instance.assets = new DLTracker(assetDlQueue, dlSize) instance.totaldlsize += dlSize*1 fulfill() }) }) } /** * Public library validation function. This function will handle the validation of libraries. * It will parse the version data, analyzing each library entry. In this analysis, it will * check to see if the local file exists and is valid. If not, it will be added to the download * queue for the 'libraries' identifier. * * @param {Object} versionData - the version data for the assets. * @param {String} basePath - the absolute file path which will be prepended to the given relative paths. * @returns {Promise.} - An empty promise to indicate the async processing has completed. */ function validateLibraries(versionData, basePath){ return new Promise(function(fulfill, reject){ const libArr = versionData.libraries const libPath = path.join(basePath, 'libraries') const libDlQueue = [] let dlSize = 0 //Check validity of each library. If the hashs don't match, download the library. async.eachLimit(libArr, 5, function(lib, cb){ if(Library.validateRules(lib.rules)){ let artifact = (lib.natives == null) ? lib.downloads.artifact : lib.downloads.classifiers[lib.natives[Library.mojangFriendlyOS()]] const libItm = new Library(lib.name, artifact.sha1, artifact.size, artifact.url, path.join(libPath, artifact.path)) if(!_validateLocal(libItm.to, 'sha1', libItm.hash)){ dlSize += (libItm.size*1) libDlQueue.push(libItm) } } cb() }, function(err){ instance.libraries = new DLTracker(libDlQueue, dlSize) instance.totaldlsize += dlSize*1 fulfill() }) }) } /** * Public miscellaneous mojang file validation function. These files will be enqueued under * the 'files' identifier. * * @param {Object} versionData - the version data for the assets. * @param {String} basePath - the absolute file path which will be prepended to the given relative paths. * @returns {Promise.} - An empty promise to indicate the async processing has completed. */ function validateMiscellaneous(versionData, basePath){ return new Promise(async function(fulfill, reject){ await validateClient(versionData, basePath) await validateLogConfig(versionData, basePath) fulfill() }) } /** * Validate client file - artifact renamed from client.jar to '{version}'.jar. * * @param {Object} versionData - the version data for the assets. * @param {String} basePath - the absolute file path which will be prepended to the given relative paths. * @param {Boolean} force - optional. If true, the asset index will be downloaded even if it exists locally. Defaults to false. * @returns {Promise.} - An empty promise to indicate the async processing has completed. */ function validateClient(versionData, basePath, force = false){ return new Promise(function(fulfill, reject){ const clientData = versionData.downloads.client const version = versionData.id const targetPath = path.join(basePath, 'versions', version) const targetFile = version + '.jar' let client = new Asset(version + ' client', clientData.sha1, clientData.size, clientData.url, path.join(targetPath, targetFile)) if(!_validateLocal(client.to, 'sha1', client.hash) || force){ instance.files.dlqueue.push(client) instance.files.dlsize += client.size*1 fulfill() } else { fulfill() } }) } /** * Validate log config. * * @param {Object} versionData - the version data for the assets. * @param {String} basePath - the absolute file path which will be prepended to the given relative paths. * @param {Boolean} force - optional. If true, the asset index will be downloaded even if it exists locally. Defaults to false. * @returns {Promise.} - An empty promise to indicate the async processing has completed. */ function validateLogConfig(versionData, basePath){ return new Promise(function(fulfill, reject){ const client = versionData.logging.client const file = client.file const targetPath = path.join(basePath, 'assets', 'log_configs') let logConfig = new Asset(file.id, file.sha1, file.size, file.url, path.join(targetPath, file.id)) if(!_validateLocal(logConfig.to, 'sha1', logConfig.hash)){ instance.files.dlqueue.push(logConfig) instance.files.dlsize += client.size*1 fulfill() } else { fulfill() } }) } function validateDistribution(serverpackid, basePath){ return new Promise(function(fulfill, reject){ _chainValidateDistributionIndex(basePath).then((value) => { let servers = value.servers let serv = null for(let i=0; i { fulfill(JSON.parse(data)) }) }) } function _parseDistroModules(modules, basePath, version){ let alist = [] let asize = 0; //This may be removed soon, considering the most efficient way to extract. let decompressqueue = [] for(let i=0; i} - A promise which resolves to Forge's version.json data. */ function loadForgeData(serverpack, basePath){ return new Promise(async function(fulfill, reject){ let distro = await _chainValidateDistributionIndex(basePath) const servers = distro.servers let serv = null for(let i=0; i} identifiers - optional. The identifiers to process and corresponding parallel async task limit. */ function processDlQueues(identifiers = [{id:'assets', limit:20}, {id:'libraries', limit:5}, {id:'files', limit:5}, {id:'forge', limit:5}]){ this.progress = 0; let win = remote.getCurrentWindow() let shouldFire = true for(let i=0; i