-
Hello, I'm still not too familiar with the project. Did the design enhancements for the landing page, but now I want to do them for the search results page. So I'm trying to add some sites to get on that page. Hopefully enough that I can get multiple pages so I can see the pagination. So I've tried to run: In there there's a comment that explains how to run this from the container: Like I mentioned I'm no Docker guru. I just know how to put containers up and build them. Look at logs and such. The instruction above, in my case, needs to be run with
|
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
Great question. First a quick clarification - the Regarding the indexing of test sites on local dev - there's a couple of options I mention in at https://github.com/searchmysite/searchmysite.net/blob/main/README.md :
There is also a 3rd option I don't mention on the README - a backup of the production search collection, in a format that can easily be imported into local dev via localhost:8983/solr/content/replication?command=restore&location=/var/solr/data/userfiles . At the moment, with approx 1,500 sites, the backup takes around 640Mb, so it isn't too difficult to work with (for comparison, the compressed wikipedia download was 32.5Gb). Not sure the best place to host it though. It is too big for GitHub. I guess I could copy from the search container to the web container and expose via the web server, like wikipedia do for their exports, although I'd need to make sure it doesn't end up increasing storage costs. |
Beta Was this translation helpful? Give feedback.
Great question.
First a quick clarification - the
src/indexing/bulkimport/wikipedia/import.sh
you ran is for the bulk load of wikipedia. It hasn't been maintained since wikipedia indexing was stopped, and now fails pretty early on. The idea was that src/indexing/bulkimport/ would contain scripts to bulk load content into the search engine directly, while src/db/bulkimport/ would contain scripts to bulk load site details into the database for the indexer to pick up and index as normal. Apologies for the confusion - I've updated the README accordingly. Fortunately the wikipedia import now fails before downloading the whole of wikipedia:-)Regarding the indexing of test sites on local dev - …