CatgirlIntelligenceAgency/run
Viktor 5bd3934d22
Merge pull request #64 from dreimolo/macos_AS_fix
Macos apple silicon fix, and slight improvements to sample downloader
2023-12-18 18:29:14 +01:00
..
dist (control, WIP) MQFSM and ProcessService are sitting in a tree 2023-07-11 17:08:43 +02:00
env (search) New frontend look 2023-12-02 17:06:40 +01:00
template/conf (conf) Change default user-agent to not associate it with the project; remove unused disks.properties file. 2023-08-01 17:34:25 +02:00
test-data Make the code run properly without WMSA_HOME set, adding missing test assets. 2023-03-05 13:47:40 +01:00
.gitignore Restructuring the git repo 2023-03-04 13:19:01 +01:00
download-samples.sh (run) Download to a temporary file to avoid corruption from aborted downloads 2023-12-18 18:28:17 +01:00
experiment-elsewhere.sh (screenshot-capture-tool) Make screenshot-capture-tool cooperate with docker 2023-11-01 16:38:55 +01:00
experiment.sh Fix experiment runner 2023-11-15 14:03:17 +01:00
readme.md (docs) Update setup instructions 2023-11-30 21:44:29 +01:00
setup.sh Merge pull request #64 from dreimolo/macos_AS_fix 2023-12-18 18:29:14 +01:00

Run

When developing locally, this directory will contain run-time data required for the search engine. In a clean check-out, it only contains the tools required to bootstrap this directory structure.

Requirements

While the system is designed to run bare metal in production, for local development, you're strongly encouraged to use docker or podman. These are a bit of a pain to install, but if you follow this guide you're on the right track.

The system requires JDK21+, and uses Java 21 preview features. Gradle complains a bit about this since it's not currently supported, but it works anyway.

Set up

To go from a clean check out of the git repo to a running search engine, follow these steps. This assumes a test deployment. For a production like setup... (TODO: write a guide for this).

You're assumed to sit in the project root the whole time.

1. Run the one-time setup

It will create the basic runtime directory structure and download some models and data that doesn't come with the git repo because git deals poorly with large binary files.

$ run/setup.sh

2. Compile the project and build docker images

$ ./gradlew docker

3. Initialize the database

Before the system can be brought online, the database needs to be initialized. To do this, bring up the database in the background, and run the flyway migration tool.

$ docker-compose up -d mariadb
$ ./gradlew flywayMigrate

4. Bring the system online.

We'll run it in the foreground in the terminal this time because it's educational to see the logs. Add -d to run in the background.

$ docker-compose up

5. You should now be able to access the system.

By default, the docker-compose file publishes the following ports:

Address Description
http://localhost:8080/ User-facing GUI
http://localhost:8081/ Operator's GUI

Note that the operator's GUI does not perform any sort of authentication.
Preferably don't expose it publicly, but if you absolutely must, use a proxy or Basic Auth to add security.

6. Download Sample Data

A script is available for downloading sample data. The script will download the data from https://downloads.marginalia.nu/ and extract it to the correct location.

The system will pick the data up automatically.

$ run/download-samples.sh l

Four sets are available:

Name Description
s Small set, 1000 domains
m Medium set, 2000 domains
l Large set, 5000 domains
xl Extra large set, 50,000 domains

Warning: The XL set is intended to provide a large amount of data for setting up a pre-production environment. It may be hard to run on a smaller machine and will on most machines take several hours to process.

The 'm' or 'l' sets are a good compromise between size and processing time and should work on most machines.

7. Process the data

Bring the system online if it isn't (see step 4), then go to the operator's GUI (see step 5).

  • Go to Node 1 -> Storage -> Crawl Data
  • Hit the toggle to set your crawl data to be active
  • Go to Actions -> Process Crawl Data -> [Trigger Reprocessing]

This will take anywhere between a few minutes to a few hours depending on which data set you downloaded. You can monitor the progress from the Overview tab.

First the CONVERTER is expected to run; this will process the data into a format that can easily be inserted into the database and index.

Next the LOADER will run; this will insert the data into the database and index.

Next the link database will repartition itself, and finally the index will be reconstructed. You can view the process of these steps in the Jobs listing.

8. Run the system

Once all this is done, you can go to the user-facing GUI (see step 5) and try a search.

Important! Use the 'No Ranking' option when running locally, since you'll very likely not have enough links for the ranking algorithm to perform well.

Experiment Runner

The script experiment.sh is a launcher for the experiment runner, which is useful when evaluating new algorithms in processing crawl data.