Your own private npm registry and backup server.
This project allows you to have your own npm registry. This server works with the necessary
npm commands just like the npmjs.org registry. You can use it to not worry about npm going down or to store your private packages. It performs much faster than npmjs.org and can even be matched with a CDN like Cloudfront to be fast globally.
Rather than trying to copy all the data in npm, this acts more like a proxy. While npm is up, it will cache package data in S3. If npm goes down, it will deliver whatever is available in the cache. This means it won't be a fully comprehensive backup of npm, but you will be able to access anything you accessed before.
The inspiration for this project comes from sinopia. This came out of a need for better cache, CDN, and general performance as well as stability of being able to run multiple instances without depending on a local filesystem.
The bulk of the data is stored in S3. You will need to set the
AWS_S3_BUCKET environment variables.
For caching, some data will be stored in
./tmp, and if
REDIS_URL is set (optional) redis will be used to cache package data.
The easiest way to set this up is with the Heroku button:
Alternatively, you can set it up by cloning this repo:
$ git clone https://github.com/dickeyxxx/elephant $ cd elephant $ npm install $ npm start
Either way, your registry is now setup and you should be able to test it by updating the packages with it:
$ npm update --registry http://urltomyregistry
See below for how to enable authorization and
How it works
Essentially the goal of the project is to quickly deliver current npm data even when npm is offline. In npm there are 2 main types of requests: package metadata and tarballs.
Package metadata mostly contains what versions of a package are available. These cannot be cached for very long since the package can be updated. By default, it is cached for 60 seconds. You can modify this with
CACHE_PACKAGE_TTL. Etags are also supported and cached to further speed up access.
The tarballs are the actual code and never change once they are uploaded (though they can be removed via unpublishing). These are downloaded one time from npmjs.org per package and version, stored in S3 and in the local tmp folder for future requests. These have a very long max-age header.
In the event npmjs.org is offline, elephant will use the most recent package metadata that was requested from npmjs.org until it comes back online.
npm commands supported
Elephant uses an htpasswd file in S3 for authentication and stores tokens in S3. To set this up, first create an htpasswd file. Then upload it to
/htpasswd in your S3 bucket. There are helpers in
./bin for this. You will need to have
AWS_S3_BUCKET environment variables set.
$ ./bin/get-htpasswd > ./htpasswd $ htpasswd -nB dickeyxxx >> ./htpasswd $ cat ./htpasswd | ./bin/put-htpasswd
Then you can login with npm. Note that the email is ignored by the server, but the CLI will force you to add one.
$ npm login --registry http://myregistry Username: dickeyxxx Password: Email: (this IS public) firstname.lastname@example.org $ npm whoami --registry http://myregistry dickeyxxx
This stores the credentials in
~/.npmrc. You can now use
npm publish to publish packages.
NOTE: Because the original use-case for having private packages was a little strange, right now you need to be authenticated to upload a private package, but once they are in the registry anyone can install them (but they would have to know the name of it). Comment on https://github.com/dickeyxxx/elephant/issues/1 if you'd like to see better functionality around this.