mirror of
https://github.com/verdaccio/verdaccio.git
synced 2024-12-23 22:27:34 -05:00
700 lines
20 KiB
JavaScript
700 lines
20 KiB
JavaScript
var assert = require('assert')
|
|
var async = require('async')
|
|
var Crypto = require('crypto')
|
|
var fs = require('fs')
|
|
var Error = require('http-errors')
|
|
var Path = require('path')
|
|
var Stream = require('readable-stream')
|
|
var URL = require('url')
|
|
var fs_storage = require('./local-fs')
|
|
var Logger = require('./logger')
|
|
var Search = require('./search')
|
|
var MyStreams = require('./streams')
|
|
var Utils = require('./utils')
|
|
var info_file = 'package.json'
|
|
|
|
//
|
|
// Implements Storage interface
|
|
// (same for storage.js, local-storage.js, up-storage.js)
|
|
//
|
|
function Storage(config) {
|
|
var self = Object.create(Storage.prototype)
|
|
self.config = config
|
|
self.logger = Logger.logger.child({ sub: 'fs' })
|
|
return self
|
|
}
|
|
|
|
// returns the minimal package file
|
|
function get_boilerplate(name) {
|
|
return {
|
|
// standard things
|
|
name: name,
|
|
versions: {},
|
|
'dist-tags': {},
|
|
|
|
// our own object
|
|
'_distfiles': {},
|
|
'_attachments': {},
|
|
'_uplinks': {},
|
|
}
|
|
}
|
|
|
|
Storage.prototype._internal_error = function(err, file, message) {
|
|
this.logger.error( { err: err, file: file }
|
|
, message + ' @{file}: @{!err.message}' )
|
|
return Error[500]()
|
|
}
|
|
|
|
Storage.prototype.add_package = function(name, info, callback) {
|
|
var storage = this.storage(name)
|
|
if (!storage) return callback( Error[404]('this package cannot be added') )
|
|
|
|
storage.create_json(info_file, get_boilerplate(name), function(err) {
|
|
if (err && err.code === 'EEXISTS') {
|
|
return callback( Error[409]('this package is already present') )
|
|
}
|
|
|
|
var latest = info['dist-tags'].latest
|
|
if (latest && info.versions[latest]) {
|
|
Search.add(info.versions[latest])
|
|
}
|
|
callback()
|
|
})
|
|
}
|
|
|
|
Storage.prototype.remove_package = function(name, callback) {
|
|
var self = this
|
|
self.logger.info( { name: name }
|
|
, 'unpublishing @{name} (all)')
|
|
|
|
var storage = self.storage(name)
|
|
if (!storage) return callback( Error[404]('no such package available') )
|
|
|
|
storage.read_json(info_file, function(err, data) {
|
|
if (err) {
|
|
if (err.code === 'ENOENT') {
|
|
return callback( Error[404]('no such package available') )
|
|
} else {
|
|
return callback(err)
|
|
}
|
|
}
|
|
self._normalize_package(data)
|
|
|
|
storage.unlink(info_file, function(err) {
|
|
if (err) return callback(err)
|
|
|
|
var files = Object.keys(data._attachments)
|
|
|
|
function unlinkNext(cb) {
|
|
if (files.length === 0) return cb()
|
|
|
|
var file = files.shift()
|
|
storage.unlink(file, function() {
|
|
unlinkNext(cb)
|
|
})
|
|
}
|
|
|
|
unlinkNext(function() {
|
|
// try to unlink the directory, but ignore errors because it can fail
|
|
storage.rmdir('.', function(err) {
|
|
callback(err)
|
|
})
|
|
})
|
|
})
|
|
})
|
|
|
|
Search.remove(name)
|
|
this.config.localList.remove(name)
|
|
}
|
|
|
|
Storage.prototype._read_create_package = function(name, callback) {
|
|
var self = this
|
|
var storage = self.storage(name)
|
|
if (!storage) {
|
|
var data = get_boilerplate(name)
|
|
self._normalize_package(data)
|
|
return callback(null, data)
|
|
}
|
|
storage.read_json(info_file, function(err, data) {
|
|
// TODO: race condition
|
|
if (err) {
|
|
if (err.code === 'ENOENT') {
|
|
// if package doesn't exist, we create it here
|
|
data = get_boilerplate(name)
|
|
} else {
|
|
return callback(self._internal_error(err, info_file, 'error reading'))
|
|
}
|
|
}
|
|
self._normalize_package(data)
|
|
callback(null, data)
|
|
})
|
|
}
|
|
|
|
// synchronize remote package info with the local one
|
|
// TODO: readfile called twice
|
|
Storage.prototype.update_versions = function(name, newdata, callback) {
|
|
var self = this
|
|
self._read_create_package(name, function(err, data) {
|
|
if (err) return callback(err)
|
|
|
|
var change = false
|
|
for (var ver in newdata.versions) {
|
|
if (data.versions[ver] == null) {
|
|
var verdata = newdata.versions[ver]
|
|
|
|
// we don't keep readmes for package versions,
|
|
// only one readme per package
|
|
delete verdata.readme
|
|
|
|
change = true
|
|
data.versions[ver] = verdata
|
|
|
|
if (verdata.dist && verdata.dist.tarball) {
|
|
var filename = URL.parse(verdata.dist.tarball).pathname.replace(/^.*\//, '')
|
|
// we do NOT overwrite any existing records
|
|
if (data._distfiles[filename] == null) {
|
|
var hash = data._distfiles[filename] = {
|
|
url: verdata.dist.tarball,
|
|
sha: verdata.dist.shasum,
|
|
}
|
|
|
|
if (verdata._sinopia_uplink) {
|
|
// if we got this information from a known registry,
|
|
// use the same protocol for the tarball
|
|
//
|
|
// see https://github.com/rlidwka/sinopia/issues/166
|
|
var tarball_url = URL.parse(hash.url)
|
|
var uplink_url = URL.parse(self.config.uplinks[verdata._sinopia_uplink].url)
|
|
if (uplink_url.host === tarball_url.host) {
|
|
tarball_url.protocol = uplink_url.protocol
|
|
hash.registry = verdata._sinopia_uplink
|
|
hash.url = URL.format(tarball_url)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
for (var tag in newdata['dist-tags']) {
|
|
if (!Array.isArray(data['dist-tags'][tag]) || data['dist-tags'][tag].length != newdata['dist-tags'][tag].length) {
|
|
// backward compat
|
|
var need_change = true
|
|
} else {
|
|
for (var i=0; i<data['dist-tags'][tag].length; i++) {
|
|
if (data['dist-tags'][tag][i] != newdata['dist-tags'][tag][i]) {
|
|
var need_change = true
|
|
break
|
|
}
|
|
}
|
|
}
|
|
|
|
if (need_change) {
|
|
change = true
|
|
data['dist-tags'][tag] = newdata['dist-tags'][tag]
|
|
}
|
|
}
|
|
for (var up in newdata._uplinks) {
|
|
var need_change = !Utils.is_object(data._uplinks[up])
|
|
|| newdata._uplinks[up].etag !== data._uplinks[up].etag
|
|
|| newdata._uplinks[up].fetched !== data._uplinks[up].fetched
|
|
|
|
if (need_change) {
|
|
change = true
|
|
data._uplinks[up] = newdata._uplinks[up]
|
|
}
|
|
}
|
|
if (newdata.readme !== data.readme) {
|
|
data.readme = newdata.readme
|
|
change = true
|
|
}
|
|
|
|
if (change) {
|
|
self.logger.debug('updating package info')
|
|
self._write_package(name, data, function(err) {
|
|
callback(err, data)
|
|
})
|
|
} else {
|
|
callback(null, data)
|
|
}
|
|
})
|
|
}
|
|
|
|
Storage.prototype.add_version = function(name, version, metadata, tag, callback) {
|
|
var self = this
|
|
self.update_package(name, function updater(data, cb) {
|
|
// keep only one readme per package
|
|
data.readme = metadata.readme
|
|
delete metadata.readme
|
|
|
|
if (data.versions[version] != null) {
|
|
return cb( Error[409]('this version already present') )
|
|
}
|
|
|
|
// if uploaded tarball has a different shasum, it's very likely that we have some kind of error
|
|
if (Utils.is_object(metadata.dist) && typeof(metadata.dist.tarball) === 'string') {
|
|
var tarball = metadata.dist.tarball.replace(/.*\//, '')
|
|
if (Utils.is_object(data._attachments[tarball])) {
|
|
if (data._attachments[tarball].shasum != null && metadata.dist.shasum != null) {
|
|
if (data._attachments[tarball].shasum != metadata.dist.shasum) {
|
|
return cb( Error[400]('shasum error, '
|
|
+ data._attachments[tarball].shasum
|
|
+ ' != ' + metadata.dist.shasum) )
|
|
}
|
|
}
|
|
|
|
data._attachments[tarball].version = version
|
|
}
|
|
}
|
|
|
|
data.versions[version] = metadata
|
|
Utils.tag_version(data, version, tag, self.config)
|
|
self.config.localList.add(name)
|
|
cb()
|
|
}, callback)
|
|
}
|
|
|
|
Storage.prototype.merge_tags = function(name, tags, callback) {
|
|
var self = this
|
|
|
|
self.update_package(name, function updater(data, cb) {
|
|
for (var t in tags) {
|
|
if (tags[t] === null) {
|
|
delete data['dist-tags'][t]
|
|
continue
|
|
}
|
|
|
|
if (data.versions[tags[t]] == null) {
|
|
return cb( Error[404]("this version doesn't exist") )
|
|
}
|
|
|
|
Utils.tag_version(data, tags[t], t, self.config)
|
|
}
|
|
cb()
|
|
}, callback)
|
|
}
|
|
|
|
Storage.prototype.replace_tags = function(name, tags, callback) {
|
|
var self = this
|
|
|
|
self.update_package(name, function updater(data, cb) {
|
|
data['dist-tags'] = {}
|
|
|
|
for (var t in tags) {
|
|
if (tags[t] === null) {
|
|
delete data['dist-tags'][t]
|
|
continue
|
|
}
|
|
|
|
if (data.versions[tags[t]] == null) {
|
|
return cb( Error[404]("this version doesn't exist") )
|
|
}
|
|
|
|
Utils.tag_version(data, tags[t], t, self.config)
|
|
}
|
|
cb()
|
|
}, callback)
|
|
}
|
|
|
|
// currently supports unpublishing only
|
|
Storage.prototype.change_package = function(name, metadata, revision, callback) {
|
|
var self = this
|
|
|
|
if (!Utils.is_object(metadata.versions) || !Utils.is_object(metadata['dist-tags'])) {
|
|
return callback( Error[422]('bad data') )
|
|
}
|
|
|
|
self.update_package(name, function updater(data, cb) {
|
|
for (var ver in data.versions) {
|
|
if (metadata.versions[ver] == null) {
|
|
self.logger.info( { name: name, version: ver }
|
|
, 'unpublishing @{name}@@{version}')
|
|
delete data.versions[ver]
|
|
|
|
for (var file in data._attachments) {
|
|
if (data._attachments[file].version === ver) {
|
|
delete data._attachments[file].version
|
|
}
|
|
}
|
|
}
|
|
}
|
|
data['dist-tags'] = metadata['dist-tags']
|
|
cb()
|
|
}, function(err) {
|
|
if (err) return callback(err)
|
|
callback()
|
|
})
|
|
}
|
|
|
|
Storage.prototype.remove_tarball = function(name, filename, revision, callback) {
|
|
assert(Utils.validate_name(filename))
|
|
var self = this
|
|
|
|
self.update_package(name, function updater(data, cb) {
|
|
if (data._attachments[filename]) {
|
|
delete data._attachments[filename]
|
|
cb()
|
|
} else {
|
|
cb(Error[404]('no such file available'))
|
|
}
|
|
}, function(err) {
|
|
if (err) return callback(err)
|
|
var storage = self.storage(name)
|
|
if (storage) storage.unlink(filename, callback)
|
|
})
|
|
}
|
|
|
|
Storage.prototype.add_tarball = function(name, filename) {
|
|
assert(Utils.validate_name(filename))
|
|
|
|
var stream = MyStreams.UploadTarballStream()
|
|
var _transform = stream._transform
|
|
var length = 0
|
|
var shasum = Crypto.createHash('sha1')
|
|
|
|
stream.abort = stream.done = function(){}
|
|
|
|
stream._transform = function(data) {
|
|
shasum.update(data)
|
|
length += data.length
|
|
_transform.apply(stream, arguments)
|
|
}
|
|
|
|
var self = this
|
|
if (name === info_file || name === '__proto__') {
|
|
process.nextTick(function() {
|
|
stream.emit('error', Error[403]("can't use this filename"))
|
|
})
|
|
return stream
|
|
}
|
|
|
|
var storage = self.storage(name)
|
|
if (!storage) {
|
|
process.nextTick(function() {
|
|
stream.emit('error', Error[404]("can't upload this package"))
|
|
})
|
|
return stream
|
|
}
|
|
|
|
var wstream = storage.write_stream(filename)
|
|
|
|
wstream.on('error', function(err) {
|
|
if (err.code === 'EEXISTS') {
|
|
stream.emit('error', Error[409]('this tarball is already present'))
|
|
} else if (err.code === 'ENOENT') {
|
|
// check if package exists to throw an appropriate message
|
|
self.get_package(name, function(_err, res) {
|
|
if (_err) {
|
|
stream.emit('error', _err)
|
|
} else {
|
|
stream.emit('error', err)
|
|
}
|
|
})
|
|
} else {
|
|
stream.emit('error', err)
|
|
}
|
|
})
|
|
|
|
wstream.on('open', function() {
|
|
// re-emitting open because it's handled in storage.js
|
|
stream.emit('open')
|
|
})
|
|
wstream.on('success', function() {
|
|
self.update_package(name, function updater(data, cb) {
|
|
data._attachments[filename] = {
|
|
shasum: shasum.digest('hex'),
|
|
}
|
|
cb()
|
|
}, function(err) {
|
|
if (err) {
|
|
stream.emit('error', err)
|
|
} else {
|
|
stream.emit('success')
|
|
}
|
|
})
|
|
})
|
|
stream.abort = function() {
|
|
wstream.abort()
|
|
}
|
|
stream.done = function() {
|
|
if (!length) {
|
|
stream.emit('error', Error[422]('refusing to accept zero-length file'))
|
|
wstream.abort()
|
|
} else {
|
|
wstream.done()
|
|
}
|
|
}
|
|
stream.pipe(wstream)
|
|
|
|
return stream
|
|
}
|
|
|
|
Storage.prototype.get_tarball = function(name, filename, callback) {
|
|
assert(Utils.validate_name(filename))
|
|
var self = this
|
|
|
|
var stream = MyStreams.ReadTarballStream()
|
|
stream.abort = function() {
|
|
if (rstream) rstream.abort()
|
|
}
|
|
|
|
var storage = self.storage(name)
|
|
if (!storage) {
|
|
process.nextTick(function() {
|
|
stream.emit('error', Error[404]('no such file available'))
|
|
})
|
|
return stream
|
|
}
|
|
|
|
var rstream = storage.read_stream(filename)
|
|
rstream.on('error', function(err) {
|
|
if (err && err.code === 'ENOENT') {
|
|
stream.emit('error', Error(404, 'no such file available'))
|
|
} else {
|
|
stream.emit('error', err)
|
|
}
|
|
})
|
|
rstream.on('content-length', function(v) {
|
|
stream.emit('content-length', v)
|
|
})
|
|
rstream.on('open', function() {
|
|
// re-emitting open because it's handled in storage.js
|
|
stream.emit('open')
|
|
rstream.pipe(stream)
|
|
})
|
|
return stream
|
|
}
|
|
|
|
Storage.prototype.get_package = function(name, options, callback) {
|
|
if (typeof(options) === 'function') callback = options, options = {}
|
|
|
|
var self = this
|
|
var storage = self.storage(name)
|
|
if (!storage) return callback( Error[404]('no such package available') )
|
|
|
|
storage.read_json(info_file, function(err, result) {
|
|
if (err) {
|
|
if (err.code === 'ENOENT') {
|
|
return callback( Error[404]('no such package available') )
|
|
} else {
|
|
return callback(self._internal_error(err, info_file, 'error reading'))
|
|
}
|
|
}
|
|
self._normalize_package(result)
|
|
callback(err, result)
|
|
})
|
|
}
|
|
|
|
// walks through each package and calls `on_package` on them
|
|
Storage.prototype._each_package = function (on_package, on_end) {
|
|
var self = this
|
|
var storages = {}
|
|
|
|
storages[self.config.storage] = true
|
|
|
|
if (self.config.packages) {
|
|
Object.keys(self.packages || {}).map(function (pkg) {
|
|
if (self.config.packages[pkg].storage) {
|
|
storages[self.config.packages[pkg].storage] = true
|
|
}
|
|
})
|
|
}
|
|
|
|
var base = Path.dirname(self.config.self_path);
|
|
|
|
async.eachSeries(Object.keys(storages), function (storage, cb) {
|
|
fs.readdir(Path.resolve(base, storage), function (err, files) {
|
|
if (err) return cb(err)
|
|
|
|
async.eachSeries(files, function (file, cb) {
|
|
if (file.match(/^@/)) {
|
|
// scoped
|
|
fs.readdir(Path.resolve(base, storage, file), function (err, files) {
|
|
if (err) return cb(err)
|
|
|
|
async.eachSeries(files, function (file2, cb) {
|
|
if (Utils.validate_name(file2)) {
|
|
on_package({
|
|
name: file + '/' + file2,
|
|
path: Path.resolve(base, storage, file, file2),
|
|
}, cb)
|
|
} else {
|
|
cb()
|
|
}
|
|
}, cb)
|
|
})
|
|
} else if (Utils.validate_name(file)) {
|
|
on_package({
|
|
name: file,
|
|
path: Path.resolve(base, storage, file)
|
|
}, cb)
|
|
} else {
|
|
cb()
|
|
}
|
|
}, cb)
|
|
})
|
|
}, on_end)
|
|
}
|
|
|
|
//
|
|
// This function allows to update the package thread-safely
|
|
//
|
|
// Arguments:
|
|
// - name - package name
|
|
// - updateFn - function(package, cb) - update function
|
|
// - callback - callback that gets invoked after it's all updated
|
|
//
|
|
// Algorithm:
|
|
// 1. lock package.json for writing
|
|
// 2. read package.json
|
|
// 3. updateFn(pkg, cb), and wait for cb
|
|
// 4. write package.json.tmp
|
|
// 5. move package.json.tmp package.json
|
|
// 6. callback(err?)
|
|
//
|
|
Storage.prototype.update_package = function(name, updateFn, _callback) {
|
|
var self = this
|
|
var storage = self.storage(name)
|
|
if (!storage) return _callback( Error[404]('no such package available') )
|
|
storage.lock_and_read_json(info_file, function(err, fd, json) {
|
|
function callback() {
|
|
var _args = arguments
|
|
if (fd) {
|
|
fs.close(fd, function(err) {
|
|
if (err) return _callback(err)
|
|
_callback.apply(null, _args)
|
|
})
|
|
} else {
|
|
_callback.apply(null, _args)
|
|
}
|
|
}
|
|
|
|
if (err) {
|
|
if (err.code === 'EAGAIN') {
|
|
return callback( Error[503]('resource temporarily unavailable') )
|
|
} else if (err.code === 'ENOENT') {
|
|
return callback( Error[404]('no such package available') )
|
|
} else {
|
|
return callback(err)
|
|
}
|
|
}
|
|
|
|
self._normalize_package(json)
|
|
updateFn(json, function(err) {
|
|
if (err) return callback(err)
|
|
|
|
self._write_package(name, json, callback)
|
|
})
|
|
})
|
|
}
|
|
|
|
Storage.prototype.search = function(startkey, options) {
|
|
var self = this
|
|
|
|
var stream = new Stream.PassThrough({ objectMode: true })
|
|
|
|
self._each_package(function on_package(item, cb) {
|
|
fs.stat(item.path, function(err, stats) {
|
|
if (err) return cb(err)
|
|
|
|
if (stats.mtime > startkey) {
|
|
self.get_package(item.name, options, function(err, data) {
|
|
if (err) return cb(err)
|
|
|
|
var versions = Utils.semver_sort(Object.keys(data.versions))
|
|
var latest = versions[versions.length - 1]
|
|
|
|
if (data.versions[latest]) {
|
|
stream.push({
|
|
name : data.versions[latest].name,
|
|
description : data.versions[latest].description,
|
|
'dist-tags' : { latest: latest },
|
|
maintainers : data.versions[latest].maintainers ||
|
|
[ data.versions[latest]._npmUser ].filter(Boolean),
|
|
author : data.versions[latest].author,
|
|
repository : data.versions[latest].repository,
|
|
readmeFilename : data.versions[latest].readmeFilename || '',
|
|
homepage : data.versions[latest].homepage,
|
|
keywords : data.versions[latest].keywords,
|
|
bugs : data.versions[latest].bugs,
|
|
license : data.versions[latest].license,
|
|
time : { modified: item.time ? new Date(item.time).toISOString() : undefined },
|
|
versions : {},
|
|
})
|
|
}
|
|
|
|
cb()
|
|
})
|
|
} else {
|
|
cb()
|
|
}
|
|
})
|
|
}, function on_end(err) {
|
|
if (err) return stream.emit('error', err)
|
|
stream.end()
|
|
})
|
|
|
|
return stream
|
|
}
|
|
|
|
Storage.prototype._normalize_package = function(pkg) {
|
|
;['versions', 'dist-tags', '_distfiles', '_attachments', '_uplinks'].forEach(function(key) {
|
|
if (!Utils.is_object(pkg[key])) pkg[key] = {}
|
|
})
|
|
if (typeof(pkg._rev) !== 'string') pkg._rev = '0-0000000000000000'
|
|
}
|
|
|
|
Storage.prototype._write_package = function(name, json, callback) {
|
|
|
|
// calculate revision a la couchdb
|
|
if (typeof(json._rev) !== 'string') json._rev = '0-0000000000000000'
|
|
var rev = json._rev.split('-')
|
|
json._rev = ((+rev[0] || 0) + 1) + '-' + Crypto.pseudoRandomBytes(8).toString('hex')
|
|
|
|
var storage = this.storage(name)
|
|
if (!storage) return callback()
|
|
storage.write_json(info_file, json, callback)
|
|
}
|
|
|
|
Storage.prototype.storage = function(package) {
|
|
var path = this.config.get_package_spec(package).storage
|
|
if (path == null) path = this.config.storage
|
|
if (path == null || path === false) {
|
|
this.logger.debug( { name: package }
|
|
, 'this package has no storage defined: @{name}' )
|
|
return null
|
|
}
|
|
return Path_Wrapper(
|
|
Path.join(
|
|
Path.resolve(Path.dirname(this.config.self_path), path),
|
|
package
|
|
)
|
|
)
|
|
}
|
|
|
|
var Path_Wrapper = (function() {
|
|
// a wrapper adding paths to fs_storage methods
|
|
function Wrapper(path) {
|
|
var self = Object.create(Wrapper.prototype)
|
|
self.path = path
|
|
return self
|
|
}
|
|
|
|
for (var i in fs_storage) {
|
|
if (fs_storage.hasOwnProperty(i)) {
|
|
Wrapper.prototype[i] = wrapper(i)
|
|
}
|
|
}
|
|
|
|
function wrapper(method) {
|
|
return function(/*...*/) {
|
|
var args = Array.prototype.slice.apply(arguments)
|
|
args[0] = Path.join(this.path, args[0] || '')
|
|
return fs_storage[method].apply(null, args)
|
|
}
|
|
}
|
|
|
|
return Wrapper
|
|
})()
|
|
|
|
module.exports = Storage
|
|
|