Your own private npm registry and backup server. Designed to be easy to set up and maintain, performant, and stable.
This project allows you to have your own npm registry. This server works with the necessary npm
commands just like the npmjs.org registry. You can use it to not worry about npm going down or to store your private packages. It performs much faster than npmjs.org and can even be matched with a CDN like Cloudfront to be fast globally.
Rather than trying to copy all the data in npm, this acts more like a proxy. While npm is up, it will cache package data locally or in S3. If npm goes down, it will deliver whatever is available in the cache. This means it won't be a fully comprehensive backup of npm, but you will be able to access anything you accessed before. This makes it easy to set up since you don't need to mirror the entire registry. Any packages previously accessed will be available.
The inspiration for this project comes from sinopia. This came out of a need for better cache, CDN, and general performance as well as stability of being able to run multiple instances without depending on a local filesystem.
This is also a 12 Factor app to make it easy to host on a PaaS like Heroku or in a custom Ansible/Chef/Puppet cluster.
The easiest way to set this up is with the Heroku button (you must use S3 with Heroku):
Alternatively, you can set it up from npm:
$ npm install -g npm-register
$ npm-register
Either way, your registry is now setup and you should be able to test it by updating the packages with it:
$ npm update --registry http://urltomyregistry
See below for how to enable authorization and npm publish
.
Use S3 for storage by setting NPM_REGISTER_STORAGE=s3
. Then set AWS_ACCESS_KEY_ID
, AWS_SECRET_ACCESS_KEY
, and AWS_S3_BUCKET
to the proper values.
Using the local filesystem is the default. You can explicitly set the storage with NPM_REGISTER_STORAGE=fs
. Select the location for the files to be stored with NPM_REGISTER_FS_DIRECTORY=/var/npm-register
. Defaults to ./tmp
.
Use GCS for storage by setting NPM_REGISTER_STORAGE=gcs
. Then set GCS_BUCKET
to the proper bucket name. Uses application default credentials.
See https://cloud.google.com/docs/authentication/production
Redis can optionally be used to cache the etags and package data. Set REDIS_URL
to activate it.
Essentially the goal of the project is to quickly deliver current npm data even when npm is offline. In npm there are 2 main types of requests: package metadata and tarballs.
Package metadata mostly contains what versions of a package are available. These cannot be cached for very long since the package can be updated. By default, it is cached for 60 seconds. You can modify this with CACHE_PACKAGE_TTL
. Etags are also supported and cached to further speed up access.
The tarballs are the actual code and never change once they are uploaded (though they can be removed via unpublishing). These are downloaded one time from npmjs.org per package and version, stored locally or in S3 for future requests. These have a very long max-age header.
In the event npmjs.org is offline, npm-register will use the most recent package metadata that was requested from npmjs.org until it comes back online.
npm-register should support most npm commands. There are some exceptions, however:
npm star
npm search
If anything else doesn't work, please submit an issue so we can fix it, or at least note the missing functionality here.
npm-register uses an htpasswd file for authentication and stores tokens in S3. To set this up, first create an htpasswd file, then upload it to /htpasswd
in your S3 bucket or your local file system:
$ aws s3 cp s3://$AWS_S3_BUCKET/htpasswd ./htpasswd
$ htpasswd -nB YOURUSERNAME >> ./htpasswd
$ aws s3 cp ./htpasswd s3://$AWS_S3_BUCKET/htpasswd
Then you can login with npm. Note that the email is ignored by the server, but the CLI will force you to add one.
$ npm login --registry http://myregistry
Username: jdxcode
Password:
Email: (this IS public) [email protected]
$ npm whoami --registry http://myregistry
jdxcode
This stores the credentials in ~/.npmrc
. You can now use npm publish
to publish packages.
By default, all write endpoints (e.g. publish, unpublish) require authentication whereas read endpoints (e.g. install) don't. This default behaviour can be changed by using NPM_REGISTER_AUTH_WRITE
and NPM_REGISTER_AUTH_READ
environment variables: use true
to enable authentication and false
to disable it.
Yarn doesn't follow HTTP redirects and so expects all URLs to be HTTPS by default. Pass
--always-https
to ignore the protocol header and return all responses in a format Yarn
understands.
To run the tests:
yarn install
yarn test
Prerequisites for running the tests locally:
The s3 bucket needs read/write/delete access. Set the following env variables:
AWS_S3_BUCKET
AWS_ACCESS_KEY_ID
AWS_SECRET_ACCESS_KEY
When not using AWS S3 but another implementation, you can use AWS_S3_PARAMS
to set
every option the official S3 client support. The Bucket is still read from the environment
variable AWS_S3_BUCKET
in this case to stay compatible with older configurations.
The following configuration worked fine for the Ceph RGW S3 implementation:
AWS_S3_BUCKET=npm-register
AWS_S3_PARAMS='{"endpoint": "http://ceph-rgw/npm-register", "s3BucketEndpoint": true, "s3DisableBodySigning": true, "accessKeyId": "<redacted>", "secretAccessKey": "<redacted>"}'
You can get a list of parameters and their meaning here: AWS SDK documentation
When running the test suite, you will need the following in ./tmp/htpasswd
and in the root of your s3 bucket:
test:$2y$05$ZhGKbrjyUbSbiMUeYeRUKOXPKzs9./NIZHsycrQkUKIj1Z2VybqdK
This sets up a test user with password 'test'.
When you have it running you should set the port number or url as the following env variable:
REDIS_URL