evazion 4c7cfc73c6 search: add new tag search parser.
Add a new tag tag search parser that supports full boolean expressions, including `and`,
`or`, and `not` operators and parenthesized subexpressions.

This is only the parser itself, not the code for converting the search into SQL. The new
parser isn't used yet for actual searches. Searches still use the old parser.

Some example syntax:

* `1girl 1boy`
* `1girl and 1boy` (same as `1girl 1boy`)
* `1girl or 1boy`
* `~1girl ~1boy` (same as `1girl or 1boy`)
* `1girl and ((blonde_hair blue_eyes) or (red_hair green_eyes))`
* `1girl ~(blonde_hair blue_eyes) ~(red_hair green_eyes)` (same as above)
* `1girl -(blonde_hair blue_eyes)`
* `*_hair *_eyes`
* `*_hair or *_eyes`
* `user:evazion or fav:evazion`
* `~user:evazion ~fav:evazion`

Rules:

AND is implicit between terms, but may be written explicitly:

* `a b c` is `a and b and c`

AND has higher precedence (binds tighter) than OR:

* `a or b and c or d` is `a or (b and c) or d`
* `a or b c or d e` is `a or (b and c) or (d and e)`

All `~` operators in the same subexpression are combined into a single OR:

* `a b ~c ~d` is `a b (c or d)`
* `~a ~b and ~c ~d` is `(a or b) (c or d)`
* `(~a ~b) (~c ~d)` is `(a or b) (c or d)`

A single `~` operator in a subexpression by itself is ignored:

* `a ~b` is `a b`
* `~a and ~b` is `a and b`, which is `a b`
* `(~a) ~b` is `a ~b`, which is `a b`

The parser is written as a backtracking recursive descent parser built on top of
StringScanner and a handful of parser combinators. The parser generates an AST, which is
then simplified using Boolean algebra to remove redundant nodes and to convert the
expression to conjunctive normal form (that is, a product of sums, or an AND of ORs).
2022-03-29 18:21:46 -05:00
2022-03-19 23:22:09 -05:00
2021-03-01 00:39:47 -06:00
2022-03-29 18:21:46 -05:00
2022-03-29 18:21:46 -05:00
2022-03-29 18:21:46 -05:00
2022-03-15 14:06:16 +01:00
2021-09-14 21:40:39 -05:00
2021-11-28 17:58:43 -06:00
2021-06-17 04:10:26 -05:00
2022-01-17 11:58:19 -06:00
2021-03-01 00:39:47 -06:00
2021-03-31 21:32:01 -05:00
2021-09-20 06:17:57 -05:00
2022-03-07 04:03:01 -06:00
2022-03-09 01:14:09 -06:00
2021-11-28 17:58:43 -06:00
2022-03-06 23:28:53 -06:00
2021-09-24 08:40:33 -05:00
2022-03-07 04:03:01 -06:00

codecov Discord

Quickstart

Run this to start a basic Danbooru instance:

curl -sSL https://raw.githubusercontent.com/danbooru/danbooru/master/bin/danbooru | sh

This will install Docker Compose and use it to start Danbooru. When it's done, Danbooru will be running at http://localhost:3000.

Alternatively, if you already have Docker Compose installed, you can just do:

wget https://raw.githubusercontent.com/danbooru/danbooru/master/docker-compose.yaml
docker-compose up

Manual Installation

Follow the INSTALL.debian script to install Danbooru.

The INSTALL.debian script is written for Debian, but can be adapted for other distributions. Danbooru has been successfully installed on Debian, Ubuntu, Fedora, Arch, and OS X. It is recommended that you use an Ubuntu-based system since Ubuntu is what is used in development and production.

See here for a guide on how set up Danbooru inside a virtual machine.

For best performance, you will need at least 256MB of RAM for PostgreSQL and Rails. The memory requirement will grow as your database gets bigger.

In production, Danbooru uses PostgreSQL 10.18, but any release later than this should work.

Troubleshooting

If your setup is not working, here are the steps I usually recommend to people:

  1. Test the database. Make sure you can connect to it using psql. Make sure the tables exist. If this fails, you need to work on correctly installing PostgreSQL, importing the initial schema, and running the migrations.

  2. Test the Rails database connection by using bin/rails console. Run Post.count to make sure Rails can connect to the database. If this fails, you need to make sure your Danbooru configuration files are correct.

  3. Test Nginx to make sure it's working correctly. You may need to debug your Nginx configuration file.

  4. Check all log files.

Services

Danboou depends on a couple of cloud services and several microservices to implement certain features.

Amazon Web Services

The following features require an Amazon AWS account:

  • Pool history
  • Post history

Google APIs

The following features require a Google Cloud account:

  • BigQuery database export

IQDB Service

IQDB integration is delegated to the IQDB service.

Archive Service

In order to access pool and post histories you will need to install and configure the Archives service.

Reportbooru Service

The following features are delegated to the Reportbooru service:

  • Post views
  • Missed searches report
  • Popular searches report

Recommender Service

Post recommendations require the Recommender service.

Description
No description provided
Readme 68 MiB
Languages
Ruby 78.3%
HTML 13.5%
JavaScript 3.5%
SCSS 2.5%
Nix 1.6%
Other 0.5%