- Matthew Woolhouse - Professor
- Jotthi Bansal - Lab instructor
- Kurt - Developer
- Michael - Developer
- Michael Mallon - IT Specialist
- Server ip address:
130.113.103.46
- Server host:
[email protected]
- Get access to the Humanities Server, ask Dr. Woolhouse or Jotthi.
- Install VPN (link) to access the server without being connected to McMaster's network.
- ssh into the server.
ssh [email protected]
.
- Download MySQL Community Edition (GPL) link.
- Change the temporary password:
Open the MySQL terminal.
mysql -u root -h 127.0.0.1 -p
For macs, there is no default MySQL config. Add a file to usr/local/mysql/etc/my.cnf
. This will skip the default password.
skip-grant-tables
Change the temporary password:
ALTER USER 'root'@'localhost' IDENTIFIED BY 'new-password';
clone the repo. get the credentials. install mysql, redis.
Rows cannot be added/deleted by the crawler if the tables in the database are not created.
- run
rsync -avz -e 'ssh -p 22' [email protected]:~/grail-dump.sql.gz ~/local_directory
.- This will download a copy of the sql data in the database currently onto your local machine.
- run
mysql -u username -p database_name < file.sql
.- In the directory that you downloaded the data into your local machine, this will populate your local database with the data and populate the database with the tables and columns.
To populate the job queue and start the crawler
- run
npm run development:master
- This initializes the dashboard and job queue.
- go to dashboard --> link
- NTS: port listed in confs
- run
npm run seed:musicbrainz:track
- This loads the queue with jobs taken from the seed files (tsv).
- run
npm run development:worker
- This plucks the jobs from the job queue and processes the jobs.
- This also inserts into the database.
To test previous crawls you will or to reset the job queue.
- run
redis-cli
- The jobs are stored in a redis database so you will need to start the redis command prompt.
- run
flushall
- This will clear the redis database and delete all the queued jobs.