Merge branch 'main' into production
This commit is contained in:
commit
21706106b5
127 changed files with 2804 additions and 895 deletions
19
.env.example
19
.env.example
|
@ -16,6 +16,11 @@ DEFAULT_LANGUAGE="English"
|
||||||
## Leave unset to allow all hosts
|
## Leave unset to allow all hosts
|
||||||
# ALLOWED_HOSTS="localhost,127.0.0.1,[::1]"
|
# ALLOWED_HOSTS="localhost,127.0.0.1,[::1]"
|
||||||
|
|
||||||
|
# Specify when the site is served from a port that is not the default
|
||||||
|
# for the protocol (80 for HTTP or 443 for HTTPS).
|
||||||
|
# Probably only necessary in development.
|
||||||
|
# PORT=1333
|
||||||
|
|
||||||
MEDIA_ROOT=images/
|
MEDIA_ROOT=images/
|
||||||
|
|
||||||
# Database configuration
|
# Database configuration
|
||||||
|
@ -71,14 +76,20 @@ ENABLE_THUMBNAIL_GENERATION=true
|
||||||
USE_S3=false
|
USE_S3=false
|
||||||
AWS_ACCESS_KEY_ID=
|
AWS_ACCESS_KEY_ID=
|
||||||
AWS_SECRET_ACCESS_KEY=
|
AWS_SECRET_ACCESS_KEY=
|
||||||
|
# seconds for signed S3 urls to expire
|
||||||
|
# this is currently only used for user export files
|
||||||
|
S3_SIGNED_URL_EXPIRY=900
|
||||||
|
|
||||||
# Commented are example values if you use a non-AWS, S3-compatible service
|
# Commented are example values if you use a non-AWS, S3-compatible service
|
||||||
# AWS S3 should work with only AWS_STORAGE_BUCKET_NAME and AWS_S3_REGION_NAME
|
# AWS S3 should work with only AWS_STORAGE_BUCKET_NAME and AWS_S3_REGION_NAME
|
||||||
# non-AWS S3-compatible services will need AWS_STORAGE_BUCKET_NAME,
|
# non-AWS S3-compatible services will need AWS_STORAGE_BUCKET_NAME,
|
||||||
# along with both AWS_S3_CUSTOM_DOMAIN and AWS_S3_ENDPOINT_URL
|
# along with both AWS_S3_CUSTOM_DOMAIN and AWS_S3_ENDPOINT_URL.
|
||||||
|
# AWS_S3_URL_PROTOCOL must end in ":" and defaults to the same protocol as
|
||||||
|
# the BookWyrm instance ("http:" or "https:", based on USE_SSL).
|
||||||
|
|
||||||
# AWS_STORAGE_BUCKET_NAME= # "example-bucket-name"
|
# AWS_STORAGE_BUCKET_NAME= # "example-bucket-name"
|
||||||
# AWS_S3_CUSTOM_DOMAIN=None # "example-bucket-name.s3.fr-par.scw.cloud"
|
# AWS_S3_CUSTOM_DOMAIN=None # "example-bucket-name.s3.fr-par.scw.cloud"
|
||||||
|
# AWS_S3_URL_PROTOCOL=None # "http:"
|
||||||
# AWS_S3_REGION_NAME=None # "fr-par"
|
# AWS_S3_REGION_NAME=None # "fr-par"
|
||||||
# AWS_S3_ENDPOINT_URL=None # "https://s3.fr-par.scw.cloud"
|
# AWS_S3_ENDPOINT_URL=None # "https://s3.fr-par.scw.cloud"
|
||||||
|
|
||||||
|
@ -133,9 +144,9 @@ HTTP_X_FORWARDED_PROTO=false
|
||||||
TWO_FACTOR_LOGIN_VALIDITY_WINDOW=2
|
TWO_FACTOR_LOGIN_VALIDITY_WINDOW=2
|
||||||
TWO_FACTOR_LOGIN_MAX_SECONDS=60
|
TWO_FACTOR_LOGIN_MAX_SECONDS=60
|
||||||
|
|
||||||
# Additional hosts to allow in the Content-Security-Policy, "self" (should be DOMAIN)
|
# Additional hosts to allow in the Content-Security-Policy, "self" (should be
|
||||||
# and AWS_S3_CUSTOM_DOMAIN (if used) are added by default.
|
# DOMAIN with optionally ":" + PORT) and AWS_S3_CUSTOM_DOMAIN (if used) are
|
||||||
# Value should be a comma-separated list of host names.
|
# added by default. Value should be a comma-separated list of host names.
|
||||||
CSP_ADDITIONAL_HOSTS=
|
CSP_ADDITIONAL_HOSTS=
|
||||||
|
|
||||||
# Time before being logged out (in seconds)
|
# Time before being logged out (in seconds)
|
||||||
|
|
78
.github/pull_request_template.md
vendored
Normal file
78
.github/pull_request_template.md
vendored
Normal file
|
@ -0,0 +1,78 @@
|
||||||
|
<!--
|
||||||
|
Thanks for contributing!
|
||||||
|
|
||||||
|
Please ensure the name of your PR is written in imperative present tense. For example:
|
||||||
|
|
||||||
|
- "fix color contrast on submit buttons"
|
||||||
|
- "add 'favourite food' value to Author model"
|
||||||
|
|
||||||
|
To check (tick) a list item, replace the space between square brackets with an x, like this:
|
||||||
|
|
||||||
|
- [x] I have checked the box
|
||||||
|
|
||||||
|
You can find more information and tips for BookWyrm contributors at https://docs.joinbookwyrm.com/contributing.html
|
||||||
|
-->
|
||||||
|
|
||||||
|
## Are you finished?
|
||||||
|
|
||||||
|
### Linters
|
||||||
|
<!--
|
||||||
|
Please run linters on your code before submitting your PR.
|
||||||
|
If you miss this step it is likely that the GitHub task runners will fail.
|
||||||
|
-->
|
||||||
|
|
||||||
|
- [ ] I have checked my code with `black`, `pylint`, and `mypy`, or `./bw-dev formatters`
|
||||||
|
|
||||||
|
### Tests
|
||||||
|
<!-- Check one -->
|
||||||
|
|
||||||
|
- [ ] My changes do not need new tests
|
||||||
|
- [ ] All tests I have added are passing
|
||||||
|
- [ ] I have written tests but need help to make them pass
|
||||||
|
- [ ] I have not written tests and need help to write them
|
||||||
|
|
||||||
|
## What type of Pull Request is this?
|
||||||
|
<!-- Check all that apply -->
|
||||||
|
|
||||||
|
- [ ] Bug Fix
|
||||||
|
- [ ] Enhancement
|
||||||
|
- [ ] Plumbing / Internals / Dependencies
|
||||||
|
- [ ] Refactor
|
||||||
|
|
||||||
|
## Does this PR change settings or dependencies, or break something?
|
||||||
|
<!-- Check all that apply -->
|
||||||
|
|
||||||
|
- [ ] This PR changes or adds default settings, configuration, or .env values
|
||||||
|
- [ ] This PR changes or adds dependencies
|
||||||
|
- [ ] This PR introduces other breaking changes
|
||||||
|
|
||||||
|
### Details of breaking or configuration changes (if any of above checked)
|
||||||
|
|
||||||
|
## Description
|
||||||
|
|
||||||
|
<!--
|
||||||
|
Describe what your pull request does here.
|
||||||
|
|
||||||
|
For pull requests that relate or close an issue, please include them
|
||||||
|
below. We like to follow [Github's guidance on linking issues to pull requests](https://docs.github.com/en/issues/tracking-your-work-with-issues/linking-a-pull-request-to-an-issue).
|
||||||
|
|
||||||
|
For example having the text: "closes #1234" would connect the current pull
|
||||||
|
request to issue 1234. And when we merge the pull request, Github will
|
||||||
|
automatically close the issue.
|
||||||
|
-->
|
||||||
|
|
||||||
|
- Related Issue #
|
||||||
|
- Closes #
|
||||||
|
|
||||||
|
## Documentation
|
||||||
|
<!--
|
||||||
|
Documentation for users, admins, and developers is an important way to keep the BookWyrm community welcoming and make Bookwyrm easy to use.
|
||||||
|
Our documentation is maintained in a separate repository at https://github.com/bookwyrm-social/documentation
|
||||||
|
-->
|
||||||
|
|
||||||
|
<!-- Check all that apply -->
|
||||||
|
|
||||||
|
- [ ] New or amended documentation will be required if this PR is merged
|
||||||
|
- [ ] I have created a matching pull request in the Documentation repository
|
||||||
|
- [ ] I intend to create a matching pull request in the Documentation repository after this PR is merged
|
||||||
|
|
26
.github/release.yml
vendored
Normal file
26
.github/release.yml
vendored
Normal file
|
@ -0,0 +1,26 @@
|
||||||
|
changelog:
|
||||||
|
exclude:
|
||||||
|
labels:
|
||||||
|
- ignore-for-release
|
||||||
|
categories:
|
||||||
|
- title: ‼️ Breaking Changes & New Settings ⚙️
|
||||||
|
labels:
|
||||||
|
- breaking-change
|
||||||
|
- config-change
|
||||||
|
- title: Updated Dependencies 🧸
|
||||||
|
labels:
|
||||||
|
- dependencies
|
||||||
|
- title: New Features 🎉
|
||||||
|
labels:
|
||||||
|
- enhancement
|
||||||
|
- title: Bug Fixes 🐛
|
||||||
|
labels:
|
||||||
|
- fix
|
||||||
|
- bug
|
||||||
|
- title: Internals/Plumbing 👩🔧
|
||||||
|
- plumbing
|
||||||
|
- tests
|
||||||
|
- deployment
|
||||||
|
- title: Other Changes
|
||||||
|
labels:
|
||||||
|
- "*"
|
3
.github/workflows/lint-frontend.yaml
vendored
3
.github/workflows/lint-frontend.yaml
vendored
|
@ -22,7 +22,8 @@ jobs:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Install modules
|
- name: Install modules
|
||||||
run: npm install stylelint stylelint-config-recommended stylelint-config-standard stylelint-order eslint
|
# run: npm install stylelint stylelint-config-recommended stylelint-config-standard stylelint-order eslint
|
||||||
|
run: npm install eslint@^8.9.0
|
||||||
|
|
||||||
# See .stylelintignore for files that are not linted.
|
# See .stylelintignore for files that are not linted.
|
||||||
# - name: Run stylelint
|
# - name: Run stylelint
|
||||||
|
|
4
.gitignore
vendored
4
.gitignore
vendored
|
@ -16,6 +16,7 @@
|
||||||
# BookWyrm
|
# BookWyrm
|
||||||
.env
|
.env
|
||||||
/images/
|
/images/
|
||||||
|
/exports/
|
||||||
/static/
|
/static/
|
||||||
bookwyrm/static/css/bookwyrm.css
|
bookwyrm/static/css/bookwyrm.css
|
||||||
bookwyrm/static/css/themes/
|
bookwyrm/static/css/themes/
|
||||||
|
@ -37,3 +38,6 @@ nginx/default.conf
|
||||||
|
|
||||||
#macOS
|
#macOS
|
||||||
**/.DS_Store
|
**/.DS_Store
|
||||||
|
|
||||||
|
# Docker
|
||||||
|
docker-compose.override.yml
|
||||||
|
|
|
@ -250,7 +250,10 @@ class ActivityObject:
|
||||||
pass
|
pass
|
||||||
data = {k: v for (k, v) in data.items() if v is not None and k not in omit}
|
data = {k: v for (k, v) in data.items() if v is not None and k not in omit}
|
||||||
if "@context" not in omit:
|
if "@context" not in omit:
|
||||||
data["@context"] = "https://www.w3.org/ns/activitystreams"
|
data["@context"] = [
|
||||||
|
"https://www.w3.org/ns/activitystreams",
|
||||||
|
{"Hashtag": "as:Hashtag"},
|
||||||
|
]
|
||||||
return data
|
return data
|
||||||
|
|
||||||
|
|
||||||
|
@ -400,11 +403,11 @@ def get_representative():
|
||||||
to sign outgoing HTTP GET requests"""
|
to sign outgoing HTTP GET requests"""
|
||||||
return models.User.objects.get_or_create(
|
return models.User.objects.get_or_create(
|
||||||
username=f"{INSTANCE_ACTOR_USERNAME}@{DOMAIN}",
|
username=f"{INSTANCE_ACTOR_USERNAME}@{DOMAIN}",
|
||||||
defaults=dict(
|
defaults={
|
||||||
email="bookwyrm@localhost",
|
"email": "bookwyrm@localhost",
|
||||||
local=True,
|
"local": True,
|
||||||
localname=INSTANCE_ACTOR_USERNAME,
|
"localname": INSTANCE_ACTOR_USERNAME,
|
||||||
),
|
},
|
||||||
)[0]
|
)[0]
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -118,9 +118,11 @@ def get_connectors() -> Iterator[abstract_connector.AbstractConnector]:
|
||||||
def get_or_create_connector(remote_id: str) -> abstract_connector.AbstractConnector:
|
def get_or_create_connector(remote_id: str) -> abstract_connector.AbstractConnector:
|
||||||
"""get the connector related to the object's server"""
|
"""get the connector related to the object's server"""
|
||||||
url = urlparse(remote_id)
|
url = urlparse(remote_id)
|
||||||
identifier = url.netloc
|
identifier = url.hostname
|
||||||
if not identifier:
|
if not identifier:
|
||||||
raise ValueError("Invalid remote id")
|
raise ValueError(f"Invalid remote id: {remote_id}")
|
||||||
|
|
||||||
|
base_url = f"{url.scheme}://{url.netloc}"
|
||||||
|
|
||||||
try:
|
try:
|
||||||
connector_info = models.Connector.objects.get(identifier=identifier)
|
connector_info = models.Connector.objects.get(identifier=identifier)
|
||||||
|
@ -128,10 +130,10 @@ def get_or_create_connector(remote_id: str) -> abstract_connector.AbstractConnec
|
||||||
connector_info = models.Connector.objects.create(
|
connector_info = models.Connector.objects.create(
|
||||||
identifier=identifier,
|
identifier=identifier,
|
||||||
connector_file="bookwyrm_connector",
|
connector_file="bookwyrm_connector",
|
||||||
base_url=f"https://{identifier}",
|
base_url=base_url,
|
||||||
books_url=f"https://{identifier}/book",
|
books_url=f"{base_url}/book",
|
||||||
covers_url=f"https://{identifier}/images/covers",
|
covers_url=f"{base_url}/images/covers",
|
||||||
search_url=f"https://{identifier}/search?q=",
|
search_url=f"{base_url}/search?q=",
|
||||||
priority=2,
|
priority=2,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -143,7 +145,9 @@ def load_more_data(connector_id: str, book_id: str) -> None:
|
||||||
"""background the work of getting all 10,000 editions of LoTR"""
|
"""background the work of getting all 10,000 editions of LoTR"""
|
||||||
connector_info = models.Connector.objects.get(id=connector_id)
|
connector_info = models.Connector.objects.get(id=connector_id)
|
||||||
connector = load_connector(connector_info)
|
connector = load_connector(connector_info)
|
||||||
book = models.Book.objects.select_subclasses().get(id=book_id)
|
book = models.Book.objects.select_subclasses().get( # type: ignore[no-untyped-call]
|
||||||
|
id=book_id
|
||||||
|
)
|
||||||
connector.expand_book_data(book)
|
connector.expand_book_data(book)
|
||||||
|
|
||||||
|
|
||||||
|
@ -154,7 +158,9 @@ def create_edition_task(
|
||||||
"""separate task for each of the 10,000 editions of LoTR"""
|
"""separate task for each of the 10,000 editions of LoTR"""
|
||||||
connector_info = models.Connector.objects.get(id=connector_id)
|
connector_info = models.Connector.objects.get(id=connector_id)
|
||||||
connector = load_connector(connector_info)
|
connector = load_connector(connector_info)
|
||||||
work = models.Work.objects.select_subclasses().get(id=work_id)
|
work = models.Work.objects.select_subclasses().get( # type: ignore[no-untyped-call]
|
||||||
|
id=work_id
|
||||||
|
)
|
||||||
connector.create_edition_from_data(work, data)
|
connector.create_edition_from_data(work, data)
|
||||||
|
|
||||||
|
|
||||||
|
@ -188,8 +194,11 @@ def raise_not_valid_url(url: str) -> None:
|
||||||
if not parsed.scheme in ["http", "https"]:
|
if not parsed.scheme in ["http", "https"]:
|
||||||
raise ConnectorException("Invalid scheme: ", url)
|
raise ConnectorException("Invalid scheme: ", url)
|
||||||
|
|
||||||
|
if not parsed.hostname:
|
||||||
|
raise ConnectorException("Hostname missing: ", url)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
ipaddress.ip_address(parsed.netloc)
|
ipaddress.ip_address(parsed.hostname)
|
||||||
raise ConnectorException("Provided url is an IP address: ", url)
|
raise ConnectorException("Provided url is an IP address: ", url)
|
||||||
except ValueError:
|
except ValueError:
|
||||||
# it's not an IP address, which is good
|
# it's not an IP address, which is good
|
||||||
|
|
|
@ -229,7 +229,7 @@ class Connector(AbstractConnector):
|
||||||
data = get_data(url)
|
data = get_data(url)
|
||||||
except ConnectorException:
|
except ConnectorException:
|
||||||
return ""
|
return ""
|
||||||
return data.get("extract", "")
|
return str(data.get("extract", ""))
|
||||||
|
|
||||||
def get_remote_id_from_model(self, obj: models.BookDataModel) -> str:
|
def get_remote_id_from_model(self, obj: models.BookDataModel) -> str:
|
||||||
"""use get_remote_id to figure out the link from a model obj"""
|
"""use get_remote_id to figure out the link from a model obj"""
|
||||||
|
|
|
@ -4,7 +4,7 @@ from django.template.loader import get_template
|
||||||
|
|
||||||
from bookwyrm import models, settings
|
from bookwyrm import models, settings
|
||||||
from bookwyrm.tasks import app, EMAIL
|
from bookwyrm.tasks import app, EMAIL
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import DOMAIN, BASE_URL
|
||||||
|
|
||||||
|
|
||||||
def email_data():
|
def email_data():
|
||||||
|
@ -14,6 +14,7 @@ def email_data():
|
||||||
"site_name": site.name,
|
"site_name": site.name,
|
||||||
"logo": site.logo_small_url,
|
"logo": site.logo_small_url,
|
||||||
"domain": DOMAIN,
|
"domain": DOMAIN,
|
||||||
|
"base_url": BASE_URL,
|
||||||
"user": None,
|
"user": None,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -26,7 +26,7 @@ class FileLinkForm(CustomForm):
|
||||||
url = cleaned_data.get("url")
|
url = cleaned_data.get("url")
|
||||||
filetype = cleaned_data.get("filetype")
|
filetype = cleaned_data.get("filetype")
|
||||||
book = cleaned_data.get("book")
|
book = cleaned_data.get("book")
|
||||||
domain = urlparse(url).netloc
|
domain = urlparse(url).hostname
|
||||||
if models.LinkDomain.objects.filter(domain=domain).exists():
|
if models.LinkDomain.objects.filter(domain=domain).exists():
|
||||||
status = models.LinkDomain.objects.get(domain=domain).status
|
status = models.LinkDomain.objects.get(domain=domain).status
|
||||||
if status == "blocked":
|
if status == "blocked":
|
||||||
|
|
|
@ -14,15 +14,10 @@ class CalibreImporter(Importer):
|
||||||
def __init__(self, *args: Any, **kwargs: Any):
|
def __init__(self, *args: Any, **kwargs: Any):
|
||||||
# Add timestamp to row_mappings_guesses for date_added to avoid
|
# Add timestamp to row_mappings_guesses for date_added to avoid
|
||||||
# integrity error
|
# integrity error
|
||||||
row_mappings_guesses = []
|
self.row_mappings_guesses = [
|
||||||
|
(field, mapping + (["timestamp"] if field == "date_added" else []))
|
||||||
for field, mapping in self.row_mappings_guesses:
|
for field, mapping in self.row_mappings_guesses
|
||||||
if field in ("date_added",):
|
]
|
||||||
row_mappings_guesses.append((field, mapping + ["timestamp"]))
|
|
||||||
else:
|
|
||||||
row_mappings_guesses.append((field, mapping))
|
|
||||||
|
|
||||||
self.row_mappings_guesses = row_mappings_guesses
|
|
||||||
super().__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
def get_shelf(self, normalized_row: dict[str, Optional[str]]) -> Optional[str]:
|
def get_shelf(self, normalized_row: dict[str, Optional[str]]) -> Optional[str]:
|
||||||
|
|
|
@ -1,13 +1,14 @@
|
||||||
""" PROCEED WITH CAUTION: uses deduplication fields to permanently
|
""" PROCEED WITH CAUTION: uses deduplication fields to permanently
|
||||||
merge book data objects """
|
merge book data objects """
|
||||||
|
|
||||||
from django.core.management.base import BaseCommand
|
from django.core.management.base import BaseCommand
|
||||||
from django.db.models import Count
|
from django.db.models import Count
|
||||||
from bookwyrm import models
|
from bookwyrm import models
|
||||||
from bookwyrm.management.merge import merge_objects
|
|
||||||
|
|
||||||
|
|
||||||
def dedupe_model(model):
|
def dedupe_model(model, dry_run=False):
|
||||||
"""combine duplicate editions and update related models"""
|
"""combine duplicate editions and update related models"""
|
||||||
|
print(f"deduplicating {model.__name__}:")
|
||||||
fields = model._meta.get_fields()
|
fields = model._meta.get_fields()
|
||||||
dedupe_fields = [
|
dedupe_fields = [
|
||||||
f for f in fields if hasattr(f, "deduplication_field") and f.deduplication_field
|
f for f in fields if hasattr(f, "deduplication_field") and f.deduplication_field
|
||||||
|
@ -16,30 +17,42 @@ def dedupe_model(model):
|
||||||
dupes = (
|
dupes = (
|
||||||
model.objects.values(field.name)
|
model.objects.values(field.name)
|
||||||
.annotate(Count(field.name))
|
.annotate(Count(field.name))
|
||||||
.filter(**{"%s__count__gt" % field.name: 1})
|
.filter(**{f"{field.name}__count__gt": 1})
|
||||||
|
.exclude(**{field.name: ""})
|
||||||
|
.exclude(**{f"{field.name}__isnull": True})
|
||||||
)
|
)
|
||||||
|
|
||||||
for dupe in dupes:
|
for dupe in dupes:
|
||||||
value = dupe[field.name]
|
value = dupe[field.name]
|
||||||
if not value or value == "":
|
|
||||||
continue
|
|
||||||
print("----------")
|
print("----------")
|
||||||
print(dupe)
|
|
||||||
objs = model.objects.filter(**{field.name: value}).order_by("id")
|
objs = model.objects.filter(**{field.name: value}).order_by("id")
|
||||||
canonical = objs.first()
|
canonical = objs.first()
|
||||||
print("keeping", canonical.remote_id)
|
action = "would merge" if dry_run else "merging"
|
||||||
|
print(
|
||||||
|
f"{action} into {model.__name__} {canonical.remote_id} based on {field.name} {value}:"
|
||||||
|
)
|
||||||
for obj in objs[1:]:
|
for obj in objs[1:]:
|
||||||
print(obj.remote_id)
|
print(f"- {obj.remote_id}")
|
||||||
merge_objects(canonical, obj)
|
absorbed_fields = obj.merge_into(canonical, dry_run=dry_run)
|
||||||
|
print(f" absorbed fields: {absorbed_fields}")
|
||||||
|
|
||||||
|
|
||||||
class Command(BaseCommand):
|
class Command(BaseCommand):
|
||||||
"""deduplicate allllll the book data models"""
|
"""deduplicate allllll the book data models"""
|
||||||
|
|
||||||
help = "merges duplicate book data"
|
help = "merges duplicate book data"
|
||||||
|
|
||||||
|
def add_arguments(self, parser):
|
||||||
|
"""add the arguments for this command"""
|
||||||
|
parser.add_argument(
|
||||||
|
"--dry_run",
|
||||||
|
action="store_true",
|
||||||
|
help="don't actually merge, only print what would happen",
|
||||||
|
)
|
||||||
|
|
||||||
# pylint: disable=no-self-use,unused-argument
|
# pylint: disable=no-self-use,unused-argument
|
||||||
def handle(self, *args, **options):
|
def handle(self, *args, **options):
|
||||||
"""run deduplications"""
|
"""run deduplications"""
|
||||||
dedupe_model(models.Edition)
|
dedupe_model(models.Edition, dry_run=options["dry_run"])
|
||||||
dedupe_model(models.Work)
|
dedupe_model(models.Work, dry_run=options["dry_run"])
|
||||||
dedupe_model(models.Author)
|
dedupe_model(models.Author, dry_run=options["dry_run"])
|
||||||
|
|
|
@ -1,50 +0,0 @@
|
||||||
from django.db.models import ManyToManyField
|
|
||||||
|
|
||||||
|
|
||||||
def update_related(canonical, obj):
|
|
||||||
"""update all the models with fk to the object being removed"""
|
|
||||||
# move related models to canonical
|
|
||||||
related_models = [
|
|
||||||
(r.remote_field.name, r.related_model) for r in canonical._meta.related_objects
|
|
||||||
]
|
|
||||||
for (related_field, related_model) in related_models:
|
|
||||||
# Skip the ManyToMany fields that aren’t auto-created. These
|
|
||||||
# should have a corresponding OneToMany field in the model for
|
|
||||||
# the linking table anyway. If we update it through that model
|
|
||||||
# instead then we won’t lose the extra fields in the linking
|
|
||||||
# table.
|
|
||||||
related_field_obj = related_model._meta.get_field(related_field)
|
|
||||||
if isinstance(related_field_obj, ManyToManyField):
|
|
||||||
through = related_field_obj.remote_field.through
|
|
||||||
if not through._meta.auto_created:
|
|
||||||
continue
|
|
||||||
related_objs = related_model.objects.filter(**{related_field: obj})
|
|
||||||
for related_obj in related_objs:
|
|
||||||
print("replacing in", related_model.__name__, related_field, related_obj.id)
|
|
||||||
try:
|
|
||||||
setattr(related_obj, related_field, canonical)
|
|
||||||
related_obj.save()
|
|
||||||
except TypeError:
|
|
||||||
getattr(related_obj, related_field).add(canonical)
|
|
||||||
getattr(related_obj, related_field).remove(obj)
|
|
||||||
|
|
||||||
|
|
||||||
def copy_data(canonical, obj):
|
|
||||||
"""try to get the most data possible"""
|
|
||||||
for data_field in obj._meta.get_fields():
|
|
||||||
if not hasattr(data_field, "activitypub_field"):
|
|
||||||
continue
|
|
||||||
data_value = getattr(obj, data_field.name)
|
|
||||||
if not data_value:
|
|
||||||
continue
|
|
||||||
if not getattr(canonical, data_field.name):
|
|
||||||
print("setting data field", data_field.name, data_value)
|
|
||||||
setattr(canonical, data_field.name, data_value)
|
|
||||||
canonical.save()
|
|
||||||
|
|
||||||
|
|
||||||
def merge_objects(canonical, obj):
|
|
||||||
copy_data(canonical, obj)
|
|
||||||
update_related(canonical, obj)
|
|
||||||
# remove the outdated entry
|
|
||||||
obj.delete()
|
|
|
@ -1,4 +1,3 @@
|
||||||
from bookwyrm.management.merge import merge_objects
|
|
||||||
from django.core.management.base import BaseCommand
|
from django.core.management.base import BaseCommand
|
||||||
|
|
||||||
|
|
||||||
|
@ -9,6 +8,11 @@ class MergeCommand(BaseCommand):
|
||||||
"""add the arguments for this command"""
|
"""add the arguments for this command"""
|
||||||
parser.add_argument("--canonical", type=int, required=True)
|
parser.add_argument("--canonical", type=int, required=True)
|
||||||
parser.add_argument("--other", type=int, required=True)
|
parser.add_argument("--other", type=int, required=True)
|
||||||
|
parser.add_argument(
|
||||||
|
"--dry_run",
|
||||||
|
action="store_true",
|
||||||
|
help="don't actually merge, only print what would happen",
|
||||||
|
)
|
||||||
|
|
||||||
# pylint: disable=no-self-use,unused-argument
|
# pylint: disable=no-self-use,unused-argument
|
||||||
def handle(self, *args, **options):
|
def handle(self, *args, **options):
|
||||||
|
@ -26,4 +30,8 @@ class MergeCommand(BaseCommand):
|
||||||
print("other book doesn’t exist!")
|
print("other book doesn’t exist!")
|
||||||
return
|
return
|
||||||
|
|
||||||
merge_objects(canonical, other)
|
absorbed_fields = other.merge_into(canonical, dry_run=options["dry_run"])
|
||||||
|
|
||||||
|
action = "would be" if options["dry_run"] else "has been"
|
||||||
|
print(f"{other.remote_id} {action} merged into {canonical.remote_id}")
|
||||||
|
print(f"absorbed fields: {absorbed_fields}")
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
""" Makes the app aware of the users timezone """
|
""" Makes the app aware of the users timezone """
|
||||||
import pytz
|
import zoneinfo
|
||||||
|
|
||||||
from django.utils import timezone
|
from django.utils import timezone
|
||||||
|
|
||||||
|
@ -12,9 +12,7 @@ class TimezoneMiddleware:
|
||||||
|
|
||||||
def __call__(self, request):
|
def __call__(self, request):
|
||||||
if request.user.is_authenticated:
|
if request.user.is_authenticated:
|
||||||
timezone.activate(pytz.timezone(request.user.preferred_timezone))
|
timezone.activate(zoneinfo.ZoneInfo(request.user.preferred_timezone))
|
||||||
else:
|
else:
|
||||||
timezone.activate(pytz.utc)
|
timezone.deactivate()
|
||||||
response = self.get_response(request)
|
return self.get_response(request)
|
||||||
timezone.deactivate()
|
|
||||||
return response
|
|
||||||
|
|
|
@ -10,6 +10,7 @@ class Migration(migrations.Migration):
|
||||||
]
|
]
|
||||||
|
|
||||||
operations = [
|
operations = [
|
||||||
|
# The new timezones are "Factory" and "localtime"
|
||||||
migrations.AlterField(
|
migrations.AlterField(
|
||||||
model_name="user",
|
model_name="user",
|
||||||
name="preferred_timezone",
|
name="preferred_timezone",
|
||||||
|
|
92
bookwyrm/migrations/0193_auto_20240128_0249.py
Normal file
92
bookwyrm/migrations/0193_auto_20240128_0249.py
Normal file
|
@ -0,0 +1,92 @@
|
||||||
|
# Generated by Django 3.2.23 on 2024-01-28 02:49
|
||||||
|
|
||||||
|
import django.core.serializers.json
|
||||||
|
from django.db import migrations, models
|
||||||
|
import django.db.models.deletion
|
||||||
|
from django.core.files.storage import storages
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0192_sitesettings_user_exports_enabled"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AddField(
|
||||||
|
model_name="bookwyrmexportjob",
|
||||||
|
name="export_json",
|
||||||
|
field=models.JSONField(
|
||||||
|
encoder=django.core.serializers.json.DjangoJSONEncoder, null=True
|
||||||
|
),
|
||||||
|
),
|
||||||
|
migrations.AddField(
|
||||||
|
model_name="bookwyrmexportjob",
|
||||||
|
name="json_completed",
|
||||||
|
field=models.BooleanField(default=False),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="bookwyrmexportjob",
|
||||||
|
name="export_data",
|
||||||
|
field=models.FileField(
|
||||||
|
null=True,
|
||||||
|
storage=storages["exports"],
|
||||||
|
upload_to="",
|
||||||
|
),
|
||||||
|
),
|
||||||
|
migrations.CreateModel(
|
||||||
|
name="AddFileToTar",
|
||||||
|
fields=[
|
||||||
|
(
|
||||||
|
"childjob_ptr",
|
||||||
|
models.OneToOneField(
|
||||||
|
auto_created=True,
|
||||||
|
on_delete=django.db.models.deletion.CASCADE,
|
||||||
|
parent_link=True,
|
||||||
|
primary_key=True,
|
||||||
|
serialize=False,
|
||||||
|
to="bookwyrm.childjob",
|
||||||
|
),
|
||||||
|
),
|
||||||
|
(
|
||||||
|
"parent_export_job",
|
||||||
|
models.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.CASCADE,
|
||||||
|
related_name="child_edition_export_jobs",
|
||||||
|
to="bookwyrm.bookwyrmexportjob",
|
||||||
|
),
|
||||||
|
),
|
||||||
|
],
|
||||||
|
options={
|
||||||
|
"abstract": False,
|
||||||
|
},
|
||||||
|
bases=("bookwyrm.childjob",),
|
||||||
|
),
|
||||||
|
migrations.CreateModel(
|
||||||
|
name="AddBookToUserExportJob",
|
||||||
|
fields=[
|
||||||
|
(
|
||||||
|
"childjob_ptr",
|
||||||
|
models.OneToOneField(
|
||||||
|
auto_created=True,
|
||||||
|
on_delete=django.db.models.deletion.CASCADE,
|
||||||
|
parent_link=True,
|
||||||
|
primary_key=True,
|
||||||
|
serialize=False,
|
||||||
|
to="bookwyrm.childjob",
|
||||||
|
),
|
||||||
|
),
|
||||||
|
(
|
||||||
|
"edition",
|
||||||
|
models.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.CASCADE,
|
||||||
|
to="bookwyrm.edition",
|
||||||
|
),
|
||||||
|
),
|
||||||
|
],
|
||||||
|
options={
|
||||||
|
"abstract": False,
|
||||||
|
},
|
||||||
|
bases=("bookwyrm.childjob",),
|
||||||
|
),
|
||||||
|
]
|
13
bookwyrm/migrations/0196_merge_20240318_1737.py
Normal file
13
bookwyrm/migrations/0196_merge_20240318_1737.py
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
# Generated by Django 3.2.23 on 2024-03-18 17:37
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0193_auto_20240128_0249"),
|
||||||
|
("bookwyrm", "0195_alter_user_preferred_language"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = []
|
13
bookwyrm/migrations/0197_merge_20240324_0235.py
Normal file
13
bookwyrm/migrations/0197_merge_20240324_0235.py
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
# Generated by Django 3.2.25 on 2024-03-24 02:35
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0196_merge_20240318_1737"),
|
||||||
|
("bookwyrm", "0196_merge_pr3134_into_main"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = []
|
48
bookwyrm/migrations/0197_mergedauthor_mergedbook.py
Normal file
48
bookwyrm/migrations/0197_mergedauthor_mergedbook.py
Normal file
|
@ -0,0 +1,48 @@
|
||||||
|
# Generated by Django 3.2.24 on 2024-02-28 21:30
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
import django.db.models.deletion
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0196_merge_pr3134_into_main"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.CreateModel(
|
||||||
|
name="MergedBook",
|
||||||
|
fields=[
|
||||||
|
("deleted_id", models.IntegerField(primary_key=True, serialize=False)),
|
||||||
|
(
|
||||||
|
"merged_into",
|
||||||
|
models.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.PROTECT,
|
||||||
|
related_name="absorbed",
|
||||||
|
to="bookwyrm.book",
|
||||||
|
),
|
||||||
|
),
|
||||||
|
],
|
||||||
|
options={
|
||||||
|
"abstract": False,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
migrations.CreateModel(
|
||||||
|
name="MergedAuthor",
|
||||||
|
fields=[
|
||||||
|
("deleted_id", models.IntegerField(primary_key=True, serialize=False)),
|
||||||
|
(
|
||||||
|
"merged_into",
|
||||||
|
models.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.PROTECT,
|
||||||
|
related_name="absorbed",
|
||||||
|
to="bookwyrm.author",
|
||||||
|
),
|
||||||
|
),
|
||||||
|
],
|
||||||
|
options={
|
||||||
|
"abstract": False,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
]
|
|
@ -0,0 +1,23 @@
|
||||||
|
# Generated by Django 3.2.25 on 2024-03-26 11:37
|
||||||
|
|
||||||
|
import bookwyrm.models.bookwyrm_export_job
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0197_merge_20240324_0235"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="bookwyrmexportjob",
|
||||||
|
name="export_data",
|
||||||
|
field=models.FileField(
|
||||||
|
null=True,
|
||||||
|
storage=bookwyrm.models.bookwyrm_export_job.select_exports_storage,
|
||||||
|
upload_to="",
|
||||||
|
),
|
||||||
|
),
|
||||||
|
]
|
|
@ -0,0 +1,70 @@
|
||||||
|
# Generated by Django 4.2.11 on 2024-03-29 19:25
|
||||||
|
|
||||||
|
import bookwyrm.models.fields
|
||||||
|
from django.conf import settings
|
||||||
|
from django.db import migrations
|
||||||
|
import django.db.models.deletion
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0198_book_search_vector_author_aliases"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="userblocks",
|
||||||
|
name="user_object",
|
||||||
|
field=bookwyrm.models.fields.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.PROTECT,
|
||||||
|
related_name="%(class)s_user_object",
|
||||||
|
to=settings.AUTH_USER_MODEL,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="userblocks",
|
||||||
|
name="user_subject",
|
||||||
|
field=bookwyrm.models.fields.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.PROTECT,
|
||||||
|
related_name="%(class)s_user_subject",
|
||||||
|
to=settings.AUTH_USER_MODEL,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="userfollowrequest",
|
||||||
|
name="user_object",
|
||||||
|
field=bookwyrm.models.fields.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.PROTECT,
|
||||||
|
related_name="%(class)s_user_object",
|
||||||
|
to=settings.AUTH_USER_MODEL,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="userfollowrequest",
|
||||||
|
name="user_subject",
|
||||||
|
field=bookwyrm.models.fields.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.PROTECT,
|
||||||
|
related_name="%(class)s_user_subject",
|
||||||
|
to=settings.AUTH_USER_MODEL,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="userfollows",
|
||||||
|
name="user_object",
|
||||||
|
field=bookwyrm.models.fields.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.PROTECT,
|
||||||
|
related_name="%(class)s_user_object",
|
||||||
|
to=settings.AUTH_USER_MODEL,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="userfollows",
|
||||||
|
name="user_subject",
|
||||||
|
field=bookwyrm.models.fields.ForeignKey(
|
||||||
|
on_delete=django.db.models.deletion.PROTECT,
|
||||||
|
related_name="%(class)s_user_subject",
|
||||||
|
to=settings.AUTH_USER_MODEL,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
]
|
13
bookwyrm/migrations/0199_merge_20240326_1217.py
Normal file
13
bookwyrm/migrations/0199_merge_20240326_1217.py
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
# Generated by Django 3.2.25 on 2024-03-26 12:17
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0198_alter_bookwyrmexportjob_export_data"),
|
||||||
|
("bookwyrm", "0198_book_search_vector_author_aliases"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = []
|
633
bookwyrm/migrations/0200_alter_user_preferred_timezone.py
Normal file
633
bookwyrm/migrations/0200_alter_user_preferred_timezone.py
Normal file
|
@ -0,0 +1,633 @@
|
||||||
|
# Generated by Django 4.2.11 on 2024-04-01 20:19
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0199_alter_userblocks_user_object_and_more"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="user",
|
||||||
|
name="preferred_timezone",
|
||||||
|
field=models.CharField(
|
||||||
|
choices=[
|
||||||
|
("Africa/Abidjan", "Africa/Abidjan"),
|
||||||
|
("Africa/Accra", "Africa/Accra"),
|
||||||
|
("Africa/Addis_Ababa", "Africa/Addis_Ababa"),
|
||||||
|
("Africa/Algiers", "Africa/Algiers"),
|
||||||
|
("Africa/Asmara", "Africa/Asmara"),
|
||||||
|
("Africa/Asmera", "Africa/Asmera"),
|
||||||
|
("Africa/Bamako", "Africa/Bamako"),
|
||||||
|
("Africa/Bangui", "Africa/Bangui"),
|
||||||
|
("Africa/Banjul", "Africa/Banjul"),
|
||||||
|
("Africa/Bissau", "Africa/Bissau"),
|
||||||
|
("Africa/Blantyre", "Africa/Blantyre"),
|
||||||
|
("Africa/Brazzaville", "Africa/Brazzaville"),
|
||||||
|
("Africa/Bujumbura", "Africa/Bujumbura"),
|
||||||
|
("Africa/Cairo", "Africa/Cairo"),
|
||||||
|
("Africa/Casablanca", "Africa/Casablanca"),
|
||||||
|
("Africa/Ceuta", "Africa/Ceuta"),
|
||||||
|
("Africa/Conakry", "Africa/Conakry"),
|
||||||
|
("Africa/Dakar", "Africa/Dakar"),
|
||||||
|
("Africa/Dar_es_Salaam", "Africa/Dar_es_Salaam"),
|
||||||
|
("Africa/Djibouti", "Africa/Djibouti"),
|
||||||
|
("Africa/Douala", "Africa/Douala"),
|
||||||
|
("Africa/El_Aaiun", "Africa/El_Aaiun"),
|
||||||
|
("Africa/Freetown", "Africa/Freetown"),
|
||||||
|
("Africa/Gaborone", "Africa/Gaborone"),
|
||||||
|
("Africa/Harare", "Africa/Harare"),
|
||||||
|
("Africa/Johannesburg", "Africa/Johannesburg"),
|
||||||
|
("Africa/Juba", "Africa/Juba"),
|
||||||
|
("Africa/Kampala", "Africa/Kampala"),
|
||||||
|
("Africa/Khartoum", "Africa/Khartoum"),
|
||||||
|
("Africa/Kigali", "Africa/Kigali"),
|
||||||
|
("Africa/Kinshasa", "Africa/Kinshasa"),
|
||||||
|
("Africa/Lagos", "Africa/Lagos"),
|
||||||
|
("Africa/Libreville", "Africa/Libreville"),
|
||||||
|
("Africa/Lome", "Africa/Lome"),
|
||||||
|
("Africa/Luanda", "Africa/Luanda"),
|
||||||
|
("Africa/Lubumbashi", "Africa/Lubumbashi"),
|
||||||
|
("Africa/Lusaka", "Africa/Lusaka"),
|
||||||
|
("Africa/Malabo", "Africa/Malabo"),
|
||||||
|
("Africa/Maputo", "Africa/Maputo"),
|
||||||
|
("Africa/Maseru", "Africa/Maseru"),
|
||||||
|
("Africa/Mbabane", "Africa/Mbabane"),
|
||||||
|
("Africa/Mogadishu", "Africa/Mogadishu"),
|
||||||
|
("Africa/Monrovia", "Africa/Monrovia"),
|
||||||
|
("Africa/Nairobi", "Africa/Nairobi"),
|
||||||
|
("Africa/Ndjamena", "Africa/Ndjamena"),
|
||||||
|
("Africa/Niamey", "Africa/Niamey"),
|
||||||
|
("Africa/Nouakchott", "Africa/Nouakchott"),
|
||||||
|
("Africa/Ouagadougou", "Africa/Ouagadougou"),
|
||||||
|
("Africa/Porto-Novo", "Africa/Porto-Novo"),
|
||||||
|
("Africa/Sao_Tome", "Africa/Sao_Tome"),
|
||||||
|
("Africa/Timbuktu", "Africa/Timbuktu"),
|
||||||
|
("Africa/Tripoli", "Africa/Tripoli"),
|
||||||
|
("Africa/Tunis", "Africa/Tunis"),
|
||||||
|
("Africa/Windhoek", "Africa/Windhoek"),
|
||||||
|
("America/Adak", "America/Adak"),
|
||||||
|
("America/Anchorage", "America/Anchorage"),
|
||||||
|
("America/Anguilla", "America/Anguilla"),
|
||||||
|
("America/Antigua", "America/Antigua"),
|
||||||
|
("America/Araguaina", "America/Araguaina"),
|
||||||
|
(
|
||||||
|
"America/Argentina/Buenos_Aires",
|
||||||
|
"America/Argentina/Buenos_Aires",
|
||||||
|
),
|
||||||
|
("America/Argentina/Catamarca", "America/Argentina/Catamarca"),
|
||||||
|
(
|
||||||
|
"America/Argentina/ComodRivadavia",
|
||||||
|
"America/Argentina/ComodRivadavia",
|
||||||
|
),
|
||||||
|
("America/Argentina/Cordoba", "America/Argentina/Cordoba"),
|
||||||
|
("America/Argentina/Jujuy", "America/Argentina/Jujuy"),
|
||||||
|
("America/Argentina/La_Rioja", "America/Argentina/La_Rioja"),
|
||||||
|
("America/Argentina/Mendoza", "America/Argentina/Mendoza"),
|
||||||
|
(
|
||||||
|
"America/Argentina/Rio_Gallegos",
|
||||||
|
"America/Argentina/Rio_Gallegos",
|
||||||
|
),
|
||||||
|
("America/Argentina/Salta", "America/Argentina/Salta"),
|
||||||
|
("America/Argentina/San_Juan", "America/Argentina/San_Juan"),
|
||||||
|
("America/Argentina/San_Luis", "America/Argentina/San_Luis"),
|
||||||
|
("America/Argentina/Tucuman", "America/Argentina/Tucuman"),
|
||||||
|
("America/Argentina/Ushuaia", "America/Argentina/Ushuaia"),
|
||||||
|
("America/Aruba", "America/Aruba"),
|
||||||
|
("America/Asuncion", "America/Asuncion"),
|
||||||
|
("America/Atikokan", "America/Atikokan"),
|
||||||
|
("America/Atka", "America/Atka"),
|
||||||
|
("America/Bahia", "America/Bahia"),
|
||||||
|
("America/Bahia_Banderas", "America/Bahia_Banderas"),
|
||||||
|
("America/Barbados", "America/Barbados"),
|
||||||
|
("America/Belem", "America/Belem"),
|
||||||
|
("America/Belize", "America/Belize"),
|
||||||
|
("America/Blanc-Sablon", "America/Blanc-Sablon"),
|
||||||
|
("America/Boa_Vista", "America/Boa_Vista"),
|
||||||
|
("America/Bogota", "America/Bogota"),
|
||||||
|
("America/Boise", "America/Boise"),
|
||||||
|
("America/Buenos_Aires", "America/Buenos_Aires"),
|
||||||
|
("America/Cambridge_Bay", "America/Cambridge_Bay"),
|
||||||
|
("America/Campo_Grande", "America/Campo_Grande"),
|
||||||
|
("America/Cancun", "America/Cancun"),
|
||||||
|
("America/Caracas", "America/Caracas"),
|
||||||
|
("America/Catamarca", "America/Catamarca"),
|
||||||
|
("America/Cayenne", "America/Cayenne"),
|
||||||
|
("America/Cayman", "America/Cayman"),
|
||||||
|
("America/Chicago", "America/Chicago"),
|
||||||
|
("America/Chihuahua", "America/Chihuahua"),
|
||||||
|
("America/Ciudad_Juarez", "America/Ciudad_Juarez"),
|
||||||
|
("America/Coral_Harbour", "America/Coral_Harbour"),
|
||||||
|
("America/Cordoba", "America/Cordoba"),
|
||||||
|
("America/Costa_Rica", "America/Costa_Rica"),
|
||||||
|
("America/Creston", "America/Creston"),
|
||||||
|
("America/Cuiaba", "America/Cuiaba"),
|
||||||
|
("America/Curacao", "America/Curacao"),
|
||||||
|
("America/Danmarkshavn", "America/Danmarkshavn"),
|
||||||
|
("America/Dawson", "America/Dawson"),
|
||||||
|
("America/Dawson_Creek", "America/Dawson_Creek"),
|
||||||
|
("America/Denver", "America/Denver"),
|
||||||
|
("America/Detroit", "America/Detroit"),
|
||||||
|
("America/Dominica", "America/Dominica"),
|
||||||
|
("America/Edmonton", "America/Edmonton"),
|
||||||
|
("America/Eirunepe", "America/Eirunepe"),
|
||||||
|
("America/El_Salvador", "America/El_Salvador"),
|
||||||
|
("America/Ensenada", "America/Ensenada"),
|
||||||
|
("America/Fort_Nelson", "America/Fort_Nelson"),
|
||||||
|
("America/Fort_Wayne", "America/Fort_Wayne"),
|
||||||
|
("America/Fortaleza", "America/Fortaleza"),
|
||||||
|
("America/Glace_Bay", "America/Glace_Bay"),
|
||||||
|
("America/Godthab", "America/Godthab"),
|
||||||
|
("America/Goose_Bay", "America/Goose_Bay"),
|
||||||
|
("America/Grand_Turk", "America/Grand_Turk"),
|
||||||
|
("America/Grenada", "America/Grenada"),
|
||||||
|
("America/Guadeloupe", "America/Guadeloupe"),
|
||||||
|
("America/Guatemala", "America/Guatemala"),
|
||||||
|
("America/Guayaquil", "America/Guayaquil"),
|
||||||
|
("America/Guyana", "America/Guyana"),
|
||||||
|
("America/Halifax", "America/Halifax"),
|
||||||
|
("America/Havana", "America/Havana"),
|
||||||
|
("America/Hermosillo", "America/Hermosillo"),
|
||||||
|
("America/Indiana/Indianapolis", "America/Indiana/Indianapolis"),
|
||||||
|
("America/Indiana/Knox", "America/Indiana/Knox"),
|
||||||
|
("America/Indiana/Marengo", "America/Indiana/Marengo"),
|
||||||
|
("America/Indiana/Petersburg", "America/Indiana/Petersburg"),
|
||||||
|
("America/Indiana/Tell_City", "America/Indiana/Tell_City"),
|
||||||
|
("America/Indiana/Vevay", "America/Indiana/Vevay"),
|
||||||
|
("America/Indiana/Vincennes", "America/Indiana/Vincennes"),
|
||||||
|
("America/Indiana/Winamac", "America/Indiana/Winamac"),
|
||||||
|
("America/Indianapolis", "America/Indianapolis"),
|
||||||
|
("America/Inuvik", "America/Inuvik"),
|
||||||
|
("America/Iqaluit", "America/Iqaluit"),
|
||||||
|
("America/Jamaica", "America/Jamaica"),
|
||||||
|
("America/Jujuy", "America/Jujuy"),
|
||||||
|
("America/Juneau", "America/Juneau"),
|
||||||
|
("America/Kentucky/Louisville", "America/Kentucky/Louisville"),
|
||||||
|
("America/Kentucky/Monticello", "America/Kentucky/Monticello"),
|
||||||
|
("America/Knox_IN", "America/Knox_IN"),
|
||||||
|
("America/Kralendijk", "America/Kralendijk"),
|
||||||
|
("America/La_Paz", "America/La_Paz"),
|
||||||
|
("America/Lima", "America/Lima"),
|
||||||
|
("America/Los_Angeles", "America/Los_Angeles"),
|
||||||
|
("America/Louisville", "America/Louisville"),
|
||||||
|
("America/Lower_Princes", "America/Lower_Princes"),
|
||||||
|
("America/Maceio", "America/Maceio"),
|
||||||
|
("America/Managua", "America/Managua"),
|
||||||
|
("America/Manaus", "America/Manaus"),
|
||||||
|
("America/Marigot", "America/Marigot"),
|
||||||
|
("America/Martinique", "America/Martinique"),
|
||||||
|
("America/Matamoros", "America/Matamoros"),
|
||||||
|
("America/Mazatlan", "America/Mazatlan"),
|
||||||
|
("America/Mendoza", "America/Mendoza"),
|
||||||
|
("America/Menominee", "America/Menominee"),
|
||||||
|
("America/Merida", "America/Merida"),
|
||||||
|
("America/Metlakatla", "America/Metlakatla"),
|
||||||
|
("America/Mexico_City", "America/Mexico_City"),
|
||||||
|
("America/Miquelon", "America/Miquelon"),
|
||||||
|
("America/Moncton", "America/Moncton"),
|
||||||
|
("America/Monterrey", "America/Monterrey"),
|
||||||
|
("America/Montevideo", "America/Montevideo"),
|
||||||
|
("America/Montreal", "America/Montreal"),
|
||||||
|
("America/Montserrat", "America/Montserrat"),
|
||||||
|
("America/Nassau", "America/Nassau"),
|
||||||
|
("America/New_York", "America/New_York"),
|
||||||
|
("America/Nipigon", "America/Nipigon"),
|
||||||
|
("America/Nome", "America/Nome"),
|
||||||
|
("America/Noronha", "America/Noronha"),
|
||||||
|
("America/North_Dakota/Beulah", "America/North_Dakota/Beulah"),
|
||||||
|
("America/North_Dakota/Center", "America/North_Dakota/Center"),
|
||||||
|
(
|
||||||
|
"America/North_Dakota/New_Salem",
|
||||||
|
"America/North_Dakota/New_Salem",
|
||||||
|
),
|
||||||
|
("America/Nuuk", "America/Nuuk"),
|
||||||
|
("America/Ojinaga", "America/Ojinaga"),
|
||||||
|
("America/Panama", "America/Panama"),
|
||||||
|
("America/Pangnirtung", "America/Pangnirtung"),
|
||||||
|
("America/Paramaribo", "America/Paramaribo"),
|
||||||
|
("America/Phoenix", "America/Phoenix"),
|
||||||
|
("America/Port-au-Prince", "America/Port-au-Prince"),
|
||||||
|
("America/Port_of_Spain", "America/Port_of_Spain"),
|
||||||
|
("America/Porto_Acre", "America/Porto_Acre"),
|
||||||
|
("America/Porto_Velho", "America/Porto_Velho"),
|
||||||
|
("America/Puerto_Rico", "America/Puerto_Rico"),
|
||||||
|
("America/Punta_Arenas", "America/Punta_Arenas"),
|
||||||
|
("America/Rainy_River", "America/Rainy_River"),
|
||||||
|
("America/Rankin_Inlet", "America/Rankin_Inlet"),
|
||||||
|
("America/Recife", "America/Recife"),
|
||||||
|
("America/Regina", "America/Regina"),
|
||||||
|
("America/Resolute", "America/Resolute"),
|
||||||
|
("America/Rio_Branco", "America/Rio_Branco"),
|
||||||
|
("America/Rosario", "America/Rosario"),
|
||||||
|
("America/Santa_Isabel", "America/Santa_Isabel"),
|
||||||
|
("America/Santarem", "America/Santarem"),
|
||||||
|
("America/Santiago", "America/Santiago"),
|
||||||
|
("America/Santo_Domingo", "America/Santo_Domingo"),
|
||||||
|
("America/Sao_Paulo", "America/Sao_Paulo"),
|
||||||
|
("America/Scoresbysund", "America/Scoresbysund"),
|
||||||
|
("America/Shiprock", "America/Shiprock"),
|
||||||
|
("America/Sitka", "America/Sitka"),
|
||||||
|
("America/St_Barthelemy", "America/St_Barthelemy"),
|
||||||
|
("America/St_Johns", "America/St_Johns"),
|
||||||
|
("America/St_Kitts", "America/St_Kitts"),
|
||||||
|
("America/St_Lucia", "America/St_Lucia"),
|
||||||
|
("America/St_Thomas", "America/St_Thomas"),
|
||||||
|
("America/St_Vincent", "America/St_Vincent"),
|
||||||
|
("America/Swift_Current", "America/Swift_Current"),
|
||||||
|
("America/Tegucigalpa", "America/Tegucigalpa"),
|
||||||
|
("America/Thule", "America/Thule"),
|
||||||
|
("America/Thunder_Bay", "America/Thunder_Bay"),
|
||||||
|
("America/Tijuana", "America/Tijuana"),
|
||||||
|
("America/Toronto", "America/Toronto"),
|
||||||
|
("America/Tortola", "America/Tortola"),
|
||||||
|
("America/Vancouver", "America/Vancouver"),
|
||||||
|
("America/Virgin", "America/Virgin"),
|
||||||
|
("America/Whitehorse", "America/Whitehorse"),
|
||||||
|
("America/Winnipeg", "America/Winnipeg"),
|
||||||
|
("America/Yakutat", "America/Yakutat"),
|
||||||
|
("America/Yellowknife", "America/Yellowknife"),
|
||||||
|
("Antarctica/Casey", "Antarctica/Casey"),
|
||||||
|
("Antarctica/Davis", "Antarctica/Davis"),
|
||||||
|
("Antarctica/DumontDUrville", "Antarctica/DumontDUrville"),
|
||||||
|
("Antarctica/Macquarie", "Antarctica/Macquarie"),
|
||||||
|
("Antarctica/Mawson", "Antarctica/Mawson"),
|
||||||
|
("Antarctica/McMurdo", "Antarctica/McMurdo"),
|
||||||
|
("Antarctica/Palmer", "Antarctica/Palmer"),
|
||||||
|
("Antarctica/Rothera", "Antarctica/Rothera"),
|
||||||
|
("Antarctica/South_Pole", "Antarctica/South_Pole"),
|
||||||
|
("Antarctica/Syowa", "Antarctica/Syowa"),
|
||||||
|
("Antarctica/Troll", "Antarctica/Troll"),
|
||||||
|
("Antarctica/Vostok", "Antarctica/Vostok"),
|
||||||
|
("Arctic/Longyearbyen", "Arctic/Longyearbyen"),
|
||||||
|
("Asia/Aden", "Asia/Aden"),
|
||||||
|
("Asia/Almaty", "Asia/Almaty"),
|
||||||
|
("Asia/Amman", "Asia/Amman"),
|
||||||
|
("Asia/Anadyr", "Asia/Anadyr"),
|
||||||
|
("Asia/Aqtau", "Asia/Aqtau"),
|
||||||
|
("Asia/Aqtobe", "Asia/Aqtobe"),
|
||||||
|
("Asia/Ashgabat", "Asia/Ashgabat"),
|
||||||
|
("Asia/Ashkhabad", "Asia/Ashkhabad"),
|
||||||
|
("Asia/Atyrau", "Asia/Atyrau"),
|
||||||
|
("Asia/Baghdad", "Asia/Baghdad"),
|
||||||
|
("Asia/Bahrain", "Asia/Bahrain"),
|
||||||
|
("Asia/Baku", "Asia/Baku"),
|
||||||
|
("Asia/Bangkok", "Asia/Bangkok"),
|
||||||
|
("Asia/Barnaul", "Asia/Barnaul"),
|
||||||
|
("Asia/Beirut", "Asia/Beirut"),
|
||||||
|
("Asia/Bishkek", "Asia/Bishkek"),
|
||||||
|
("Asia/Brunei", "Asia/Brunei"),
|
||||||
|
("Asia/Calcutta", "Asia/Calcutta"),
|
||||||
|
("Asia/Chita", "Asia/Chita"),
|
||||||
|
("Asia/Choibalsan", "Asia/Choibalsan"),
|
||||||
|
("Asia/Chongqing", "Asia/Chongqing"),
|
||||||
|
("Asia/Chungking", "Asia/Chungking"),
|
||||||
|
("Asia/Colombo", "Asia/Colombo"),
|
||||||
|
("Asia/Dacca", "Asia/Dacca"),
|
||||||
|
("Asia/Damascus", "Asia/Damascus"),
|
||||||
|
("Asia/Dhaka", "Asia/Dhaka"),
|
||||||
|
("Asia/Dili", "Asia/Dili"),
|
||||||
|
("Asia/Dubai", "Asia/Dubai"),
|
||||||
|
("Asia/Dushanbe", "Asia/Dushanbe"),
|
||||||
|
("Asia/Famagusta", "Asia/Famagusta"),
|
||||||
|
("Asia/Gaza", "Asia/Gaza"),
|
||||||
|
("Asia/Harbin", "Asia/Harbin"),
|
||||||
|
("Asia/Hebron", "Asia/Hebron"),
|
||||||
|
("Asia/Ho_Chi_Minh", "Asia/Ho_Chi_Minh"),
|
||||||
|
("Asia/Hong_Kong", "Asia/Hong_Kong"),
|
||||||
|
("Asia/Hovd", "Asia/Hovd"),
|
||||||
|
("Asia/Irkutsk", "Asia/Irkutsk"),
|
||||||
|
("Asia/Istanbul", "Asia/Istanbul"),
|
||||||
|
("Asia/Jakarta", "Asia/Jakarta"),
|
||||||
|
("Asia/Jayapura", "Asia/Jayapura"),
|
||||||
|
("Asia/Jerusalem", "Asia/Jerusalem"),
|
||||||
|
("Asia/Kabul", "Asia/Kabul"),
|
||||||
|
("Asia/Kamchatka", "Asia/Kamchatka"),
|
||||||
|
("Asia/Karachi", "Asia/Karachi"),
|
||||||
|
("Asia/Kashgar", "Asia/Kashgar"),
|
||||||
|
("Asia/Kathmandu", "Asia/Kathmandu"),
|
||||||
|
("Asia/Katmandu", "Asia/Katmandu"),
|
||||||
|
("Asia/Khandyga", "Asia/Khandyga"),
|
||||||
|
("Asia/Kolkata", "Asia/Kolkata"),
|
||||||
|
("Asia/Krasnoyarsk", "Asia/Krasnoyarsk"),
|
||||||
|
("Asia/Kuala_Lumpur", "Asia/Kuala_Lumpur"),
|
||||||
|
("Asia/Kuching", "Asia/Kuching"),
|
||||||
|
("Asia/Kuwait", "Asia/Kuwait"),
|
||||||
|
("Asia/Macao", "Asia/Macao"),
|
||||||
|
("Asia/Macau", "Asia/Macau"),
|
||||||
|
("Asia/Magadan", "Asia/Magadan"),
|
||||||
|
("Asia/Makassar", "Asia/Makassar"),
|
||||||
|
("Asia/Manila", "Asia/Manila"),
|
||||||
|
("Asia/Muscat", "Asia/Muscat"),
|
||||||
|
("Asia/Nicosia", "Asia/Nicosia"),
|
||||||
|
("Asia/Novokuznetsk", "Asia/Novokuznetsk"),
|
||||||
|
("Asia/Novosibirsk", "Asia/Novosibirsk"),
|
||||||
|
("Asia/Omsk", "Asia/Omsk"),
|
||||||
|
("Asia/Oral", "Asia/Oral"),
|
||||||
|
("Asia/Phnom_Penh", "Asia/Phnom_Penh"),
|
||||||
|
("Asia/Pontianak", "Asia/Pontianak"),
|
||||||
|
("Asia/Pyongyang", "Asia/Pyongyang"),
|
||||||
|
("Asia/Qatar", "Asia/Qatar"),
|
||||||
|
("Asia/Qostanay", "Asia/Qostanay"),
|
||||||
|
("Asia/Qyzylorda", "Asia/Qyzylorda"),
|
||||||
|
("Asia/Rangoon", "Asia/Rangoon"),
|
||||||
|
("Asia/Riyadh", "Asia/Riyadh"),
|
||||||
|
("Asia/Saigon", "Asia/Saigon"),
|
||||||
|
("Asia/Sakhalin", "Asia/Sakhalin"),
|
||||||
|
("Asia/Samarkand", "Asia/Samarkand"),
|
||||||
|
("Asia/Seoul", "Asia/Seoul"),
|
||||||
|
("Asia/Shanghai", "Asia/Shanghai"),
|
||||||
|
("Asia/Singapore", "Asia/Singapore"),
|
||||||
|
("Asia/Srednekolymsk", "Asia/Srednekolymsk"),
|
||||||
|
("Asia/Taipei", "Asia/Taipei"),
|
||||||
|
("Asia/Tashkent", "Asia/Tashkent"),
|
||||||
|
("Asia/Tbilisi", "Asia/Tbilisi"),
|
||||||
|
("Asia/Tehran", "Asia/Tehran"),
|
||||||
|
("Asia/Tel_Aviv", "Asia/Tel_Aviv"),
|
||||||
|
("Asia/Thimbu", "Asia/Thimbu"),
|
||||||
|
("Asia/Thimphu", "Asia/Thimphu"),
|
||||||
|
("Asia/Tokyo", "Asia/Tokyo"),
|
||||||
|
("Asia/Tomsk", "Asia/Tomsk"),
|
||||||
|
("Asia/Ujung_Pandang", "Asia/Ujung_Pandang"),
|
||||||
|
("Asia/Ulaanbaatar", "Asia/Ulaanbaatar"),
|
||||||
|
("Asia/Ulan_Bator", "Asia/Ulan_Bator"),
|
||||||
|
("Asia/Urumqi", "Asia/Urumqi"),
|
||||||
|
("Asia/Ust-Nera", "Asia/Ust-Nera"),
|
||||||
|
("Asia/Vientiane", "Asia/Vientiane"),
|
||||||
|
("Asia/Vladivostok", "Asia/Vladivostok"),
|
||||||
|
("Asia/Yakutsk", "Asia/Yakutsk"),
|
||||||
|
("Asia/Yangon", "Asia/Yangon"),
|
||||||
|
("Asia/Yekaterinburg", "Asia/Yekaterinburg"),
|
||||||
|
("Asia/Yerevan", "Asia/Yerevan"),
|
||||||
|
("Atlantic/Azores", "Atlantic/Azores"),
|
||||||
|
("Atlantic/Bermuda", "Atlantic/Bermuda"),
|
||||||
|
("Atlantic/Canary", "Atlantic/Canary"),
|
||||||
|
("Atlantic/Cape_Verde", "Atlantic/Cape_Verde"),
|
||||||
|
("Atlantic/Faeroe", "Atlantic/Faeroe"),
|
||||||
|
("Atlantic/Faroe", "Atlantic/Faroe"),
|
||||||
|
("Atlantic/Jan_Mayen", "Atlantic/Jan_Mayen"),
|
||||||
|
("Atlantic/Madeira", "Atlantic/Madeira"),
|
||||||
|
("Atlantic/Reykjavik", "Atlantic/Reykjavik"),
|
||||||
|
("Atlantic/South_Georgia", "Atlantic/South_Georgia"),
|
||||||
|
("Atlantic/St_Helena", "Atlantic/St_Helena"),
|
||||||
|
("Atlantic/Stanley", "Atlantic/Stanley"),
|
||||||
|
("Australia/ACT", "Australia/ACT"),
|
||||||
|
("Australia/Adelaide", "Australia/Adelaide"),
|
||||||
|
("Australia/Brisbane", "Australia/Brisbane"),
|
||||||
|
("Australia/Broken_Hill", "Australia/Broken_Hill"),
|
||||||
|
("Australia/Canberra", "Australia/Canberra"),
|
||||||
|
("Australia/Currie", "Australia/Currie"),
|
||||||
|
("Australia/Darwin", "Australia/Darwin"),
|
||||||
|
("Australia/Eucla", "Australia/Eucla"),
|
||||||
|
("Australia/Hobart", "Australia/Hobart"),
|
||||||
|
("Australia/LHI", "Australia/LHI"),
|
||||||
|
("Australia/Lindeman", "Australia/Lindeman"),
|
||||||
|
("Australia/Lord_Howe", "Australia/Lord_Howe"),
|
||||||
|
("Australia/Melbourne", "Australia/Melbourne"),
|
||||||
|
("Australia/NSW", "Australia/NSW"),
|
||||||
|
("Australia/North", "Australia/North"),
|
||||||
|
("Australia/Perth", "Australia/Perth"),
|
||||||
|
("Australia/Queensland", "Australia/Queensland"),
|
||||||
|
("Australia/South", "Australia/South"),
|
||||||
|
("Australia/Sydney", "Australia/Sydney"),
|
||||||
|
("Australia/Tasmania", "Australia/Tasmania"),
|
||||||
|
("Australia/Victoria", "Australia/Victoria"),
|
||||||
|
("Australia/West", "Australia/West"),
|
||||||
|
("Australia/Yancowinna", "Australia/Yancowinna"),
|
||||||
|
("Brazil/Acre", "Brazil/Acre"),
|
||||||
|
("Brazil/DeNoronha", "Brazil/DeNoronha"),
|
||||||
|
("Brazil/East", "Brazil/East"),
|
||||||
|
("Brazil/West", "Brazil/West"),
|
||||||
|
("CET", "CET"),
|
||||||
|
("CST6CDT", "CST6CDT"),
|
||||||
|
("Canada/Atlantic", "Canada/Atlantic"),
|
||||||
|
("Canada/Central", "Canada/Central"),
|
||||||
|
("Canada/Eastern", "Canada/Eastern"),
|
||||||
|
("Canada/Mountain", "Canada/Mountain"),
|
||||||
|
("Canada/Newfoundland", "Canada/Newfoundland"),
|
||||||
|
("Canada/Pacific", "Canada/Pacific"),
|
||||||
|
("Canada/Saskatchewan", "Canada/Saskatchewan"),
|
||||||
|
("Canada/Yukon", "Canada/Yukon"),
|
||||||
|
("Chile/Continental", "Chile/Continental"),
|
||||||
|
("Chile/EasterIsland", "Chile/EasterIsland"),
|
||||||
|
("Cuba", "Cuba"),
|
||||||
|
("EET", "EET"),
|
||||||
|
("EST", "EST"),
|
||||||
|
("EST5EDT", "EST5EDT"),
|
||||||
|
("Egypt", "Egypt"),
|
||||||
|
("Eire", "Eire"),
|
||||||
|
("Etc/GMT", "Etc/GMT"),
|
||||||
|
("Etc/GMT+0", "Etc/GMT+0"),
|
||||||
|
("Etc/GMT+1", "Etc/GMT+1"),
|
||||||
|
("Etc/GMT+10", "Etc/GMT+10"),
|
||||||
|
("Etc/GMT+11", "Etc/GMT+11"),
|
||||||
|
("Etc/GMT+12", "Etc/GMT+12"),
|
||||||
|
("Etc/GMT+2", "Etc/GMT+2"),
|
||||||
|
("Etc/GMT+3", "Etc/GMT+3"),
|
||||||
|
("Etc/GMT+4", "Etc/GMT+4"),
|
||||||
|
("Etc/GMT+5", "Etc/GMT+5"),
|
||||||
|
("Etc/GMT+6", "Etc/GMT+6"),
|
||||||
|
("Etc/GMT+7", "Etc/GMT+7"),
|
||||||
|
("Etc/GMT+8", "Etc/GMT+8"),
|
||||||
|
("Etc/GMT+9", "Etc/GMT+9"),
|
||||||
|
("Etc/GMT-0", "Etc/GMT-0"),
|
||||||
|
("Etc/GMT-1", "Etc/GMT-1"),
|
||||||
|
("Etc/GMT-10", "Etc/GMT-10"),
|
||||||
|
("Etc/GMT-11", "Etc/GMT-11"),
|
||||||
|
("Etc/GMT-12", "Etc/GMT-12"),
|
||||||
|
("Etc/GMT-13", "Etc/GMT-13"),
|
||||||
|
("Etc/GMT-14", "Etc/GMT-14"),
|
||||||
|
("Etc/GMT-2", "Etc/GMT-2"),
|
||||||
|
("Etc/GMT-3", "Etc/GMT-3"),
|
||||||
|
("Etc/GMT-4", "Etc/GMT-4"),
|
||||||
|
("Etc/GMT-5", "Etc/GMT-5"),
|
||||||
|
("Etc/GMT-6", "Etc/GMT-6"),
|
||||||
|
("Etc/GMT-7", "Etc/GMT-7"),
|
||||||
|
("Etc/GMT-8", "Etc/GMT-8"),
|
||||||
|
("Etc/GMT-9", "Etc/GMT-9"),
|
||||||
|
("Etc/GMT0", "Etc/GMT0"),
|
||||||
|
("Etc/Greenwich", "Etc/Greenwich"),
|
||||||
|
("Etc/UCT", "Etc/UCT"),
|
||||||
|
("Etc/UTC", "Etc/UTC"),
|
||||||
|
("Etc/Universal", "Etc/Universal"),
|
||||||
|
("Etc/Zulu", "Etc/Zulu"),
|
||||||
|
("Europe/Amsterdam", "Europe/Amsterdam"),
|
||||||
|
("Europe/Andorra", "Europe/Andorra"),
|
||||||
|
("Europe/Astrakhan", "Europe/Astrakhan"),
|
||||||
|
("Europe/Athens", "Europe/Athens"),
|
||||||
|
("Europe/Belfast", "Europe/Belfast"),
|
||||||
|
("Europe/Belgrade", "Europe/Belgrade"),
|
||||||
|
("Europe/Berlin", "Europe/Berlin"),
|
||||||
|
("Europe/Bratislava", "Europe/Bratislava"),
|
||||||
|
("Europe/Brussels", "Europe/Brussels"),
|
||||||
|
("Europe/Bucharest", "Europe/Bucharest"),
|
||||||
|
("Europe/Budapest", "Europe/Budapest"),
|
||||||
|
("Europe/Busingen", "Europe/Busingen"),
|
||||||
|
("Europe/Chisinau", "Europe/Chisinau"),
|
||||||
|
("Europe/Copenhagen", "Europe/Copenhagen"),
|
||||||
|
("Europe/Dublin", "Europe/Dublin"),
|
||||||
|
("Europe/Gibraltar", "Europe/Gibraltar"),
|
||||||
|
("Europe/Guernsey", "Europe/Guernsey"),
|
||||||
|
("Europe/Helsinki", "Europe/Helsinki"),
|
||||||
|
("Europe/Isle_of_Man", "Europe/Isle_of_Man"),
|
||||||
|
("Europe/Istanbul", "Europe/Istanbul"),
|
||||||
|
("Europe/Jersey", "Europe/Jersey"),
|
||||||
|
("Europe/Kaliningrad", "Europe/Kaliningrad"),
|
||||||
|
("Europe/Kiev", "Europe/Kiev"),
|
||||||
|
("Europe/Kirov", "Europe/Kirov"),
|
||||||
|
("Europe/Kyiv", "Europe/Kyiv"),
|
||||||
|
("Europe/Lisbon", "Europe/Lisbon"),
|
||||||
|
("Europe/Ljubljana", "Europe/Ljubljana"),
|
||||||
|
("Europe/London", "Europe/London"),
|
||||||
|
("Europe/Luxembourg", "Europe/Luxembourg"),
|
||||||
|
("Europe/Madrid", "Europe/Madrid"),
|
||||||
|
("Europe/Malta", "Europe/Malta"),
|
||||||
|
("Europe/Mariehamn", "Europe/Mariehamn"),
|
||||||
|
("Europe/Minsk", "Europe/Minsk"),
|
||||||
|
("Europe/Monaco", "Europe/Monaco"),
|
||||||
|
("Europe/Moscow", "Europe/Moscow"),
|
||||||
|
("Europe/Nicosia", "Europe/Nicosia"),
|
||||||
|
("Europe/Oslo", "Europe/Oslo"),
|
||||||
|
("Europe/Paris", "Europe/Paris"),
|
||||||
|
("Europe/Podgorica", "Europe/Podgorica"),
|
||||||
|
("Europe/Prague", "Europe/Prague"),
|
||||||
|
("Europe/Riga", "Europe/Riga"),
|
||||||
|
("Europe/Rome", "Europe/Rome"),
|
||||||
|
("Europe/Samara", "Europe/Samara"),
|
||||||
|
("Europe/San_Marino", "Europe/San_Marino"),
|
||||||
|
("Europe/Sarajevo", "Europe/Sarajevo"),
|
||||||
|
("Europe/Saratov", "Europe/Saratov"),
|
||||||
|
("Europe/Simferopol", "Europe/Simferopol"),
|
||||||
|
("Europe/Skopje", "Europe/Skopje"),
|
||||||
|
("Europe/Sofia", "Europe/Sofia"),
|
||||||
|
("Europe/Stockholm", "Europe/Stockholm"),
|
||||||
|
("Europe/Tallinn", "Europe/Tallinn"),
|
||||||
|
("Europe/Tirane", "Europe/Tirane"),
|
||||||
|
("Europe/Tiraspol", "Europe/Tiraspol"),
|
||||||
|
("Europe/Ulyanovsk", "Europe/Ulyanovsk"),
|
||||||
|
("Europe/Uzhgorod", "Europe/Uzhgorod"),
|
||||||
|
("Europe/Vaduz", "Europe/Vaduz"),
|
||||||
|
("Europe/Vatican", "Europe/Vatican"),
|
||||||
|
("Europe/Vienna", "Europe/Vienna"),
|
||||||
|
("Europe/Vilnius", "Europe/Vilnius"),
|
||||||
|
("Europe/Volgograd", "Europe/Volgograd"),
|
||||||
|
("Europe/Warsaw", "Europe/Warsaw"),
|
||||||
|
("Europe/Zagreb", "Europe/Zagreb"),
|
||||||
|
("Europe/Zaporozhye", "Europe/Zaporozhye"),
|
||||||
|
("Europe/Zurich", "Europe/Zurich"),
|
||||||
|
("Factory", "Factory"),
|
||||||
|
("GB", "GB"),
|
||||||
|
("GB-Eire", "GB-Eire"),
|
||||||
|
("GMT", "GMT"),
|
||||||
|
("GMT+0", "GMT+0"),
|
||||||
|
("GMT-0", "GMT-0"),
|
||||||
|
("GMT0", "GMT0"),
|
||||||
|
("Greenwich", "Greenwich"),
|
||||||
|
("HST", "HST"),
|
||||||
|
("Hongkong", "Hongkong"),
|
||||||
|
("Iceland", "Iceland"),
|
||||||
|
("Indian/Antananarivo", "Indian/Antananarivo"),
|
||||||
|
("Indian/Chagos", "Indian/Chagos"),
|
||||||
|
("Indian/Christmas", "Indian/Christmas"),
|
||||||
|
("Indian/Cocos", "Indian/Cocos"),
|
||||||
|
("Indian/Comoro", "Indian/Comoro"),
|
||||||
|
("Indian/Kerguelen", "Indian/Kerguelen"),
|
||||||
|
("Indian/Mahe", "Indian/Mahe"),
|
||||||
|
("Indian/Maldives", "Indian/Maldives"),
|
||||||
|
("Indian/Mauritius", "Indian/Mauritius"),
|
||||||
|
("Indian/Mayotte", "Indian/Mayotte"),
|
||||||
|
("Indian/Reunion", "Indian/Reunion"),
|
||||||
|
("Iran", "Iran"),
|
||||||
|
("Israel", "Israel"),
|
||||||
|
("Jamaica", "Jamaica"),
|
||||||
|
("Japan", "Japan"),
|
||||||
|
("Kwajalein", "Kwajalein"),
|
||||||
|
("Libya", "Libya"),
|
||||||
|
("MET", "MET"),
|
||||||
|
("MST", "MST"),
|
||||||
|
("MST7MDT", "MST7MDT"),
|
||||||
|
("Mexico/BajaNorte", "Mexico/BajaNorte"),
|
||||||
|
("Mexico/BajaSur", "Mexico/BajaSur"),
|
||||||
|
("Mexico/General", "Mexico/General"),
|
||||||
|
("NZ", "NZ"),
|
||||||
|
("NZ-CHAT", "NZ-CHAT"),
|
||||||
|
("Navajo", "Navajo"),
|
||||||
|
("PRC", "PRC"),
|
||||||
|
("PST8PDT", "PST8PDT"),
|
||||||
|
("Pacific/Apia", "Pacific/Apia"),
|
||||||
|
("Pacific/Auckland", "Pacific/Auckland"),
|
||||||
|
("Pacific/Bougainville", "Pacific/Bougainville"),
|
||||||
|
("Pacific/Chatham", "Pacific/Chatham"),
|
||||||
|
("Pacific/Chuuk", "Pacific/Chuuk"),
|
||||||
|
("Pacific/Easter", "Pacific/Easter"),
|
||||||
|
("Pacific/Efate", "Pacific/Efate"),
|
||||||
|
("Pacific/Enderbury", "Pacific/Enderbury"),
|
||||||
|
("Pacific/Fakaofo", "Pacific/Fakaofo"),
|
||||||
|
("Pacific/Fiji", "Pacific/Fiji"),
|
||||||
|
("Pacific/Funafuti", "Pacific/Funafuti"),
|
||||||
|
("Pacific/Galapagos", "Pacific/Galapagos"),
|
||||||
|
("Pacific/Gambier", "Pacific/Gambier"),
|
||||||
|
("Pacific/Guadalcanal", "Pacific/Guadalcanal"),
|
||||||
|
("Pacific/Guam", "Pacific/Guam"),
|
||||||
|
("Pacific/Honolulu", "Pacific/Honolulu"),
|
||||||
|
("Pacific/Johnston", "Pacific/Johnston"),
|
||||||
|
("Pacific/Kanton", "Pacific/Kanton"),
|
||||||
|
("Pacific/Kiritimati", "Pacific/Kiritimati"),
|
||||||
|
("Pacific/Kosrae", "Pacific/Kosrae"),
|
||||||
|
("Pacific/Kwajalein", "Pacific/Kwajalein"),
|
||||||
|
("Pacific/Majuro", "Pacific/Majuro"),
|
||||||
|
("Pacific/Marquesas", "Pacific/Marquesas"),
|
||||||
|
("Pacific/Midway", "Pacific/Midway"),
|
||||||
|
("Pacific/Nauru", "Pacific/Nauru"),
|
||||||
|
("Pacific/Niue", "Pacific/Niue"),
|
||||||
|
("Pacific/Norfolk", "Pacific/Norfolk"),
|
||||||
|
("Pacific/Noumea", "Pacific/Noumea"),
|
||||||
|
("Pacific/Pago_Pago", "Pacific/Pago_Pago"),
|
||||||
|
("Pacific/Palau", "Pacific/Palau"),
|
||||||
|
("Pacific/Pitcairn", "Pacific/Pitcairn"),
|
||||||
|
("Pacific/Pohnpei", "Pacific/Pohnpei"),
|
||||||
|
("Pacific/Ponape", "Pacific/Ponape"),
|
||||||
|
("Pacific/Port_Moresby", "Pacific/Port_Moresby"),
|
||||||
|
("Pacific/Rarotonga", "Pacific/Rarotonga"),
|
||||||
|
("Pacific/Saipan", "Pacific/Saipan"),
|
||||||
|
("Pacific/Samoa", "Pacific/Samoa"),
|
||||||
|
("Pacific/Tahiti", "Pacific/Tahiti"),
|
||||||
|
("Pacific/Tarawa", "Pacific/Tarawa"),
|
||||||
|
("Pacific/Tongatapu", "Pacific/Tongatapu"),
|
||||||
|
("Pacific/Truk", "Pacific/Truk"),
|
||||||
|
("Pacific/Wake", "Pacific/Wake"),
|
||||||
|
("Pacific/Wallis", "Pacific/Wallis"),
|
||||||
|
("Pacific/Yap", "Pacific/Yap"),
|
||||||
|
("Poland", "Poland"),
|
||||||
|
("Portugal", "Portugal"),
|
||||||
|
("ROC", "ROC"),
|
||||||
|
("ROK", "ROK"),
|
||||||
|
("Singapore", "Singapore"),
|
||||||
|
("Turkey", "Turkey"),
|
||||||
|
("UCT", "UCT"),
|
||||||
|
("US/Alaska", "US/Alaska"),
|
||||||
|
("US/Aleutian", "US/Aleutian"),
|
||||||
|
("US/Arizona", "US/Arizona"),
|
||||||
|
("US/Central", "US/Central"),
|
||||||
|
("US/East-Indiana", "US/East-Indiana"),
|
||||||
|
("US/Eastern", "US/Eastern"),
|
||||||
|
("US/Hawaii", "US/Hawaii"),
|
||||||
|
("US/Indiana-Starke", "US/Indiana-Starke"),
|
||||||
|
("US/Michigan", "US/Michigan"),
|
||||||
|
("US/Mountain", "US/Mountain"),
|
||||||
|
("US/Pacific", "US/Pacific"),
|
||||||
|
("US/Samoa", "US/Samoa"),
|
||||||
|
("UTC", "UTC"),
|
||||||
|
("Universal", "Universal"),
|
||||||
|
("W-SU", "W-SU"),
|
||||||
|
("WET", "WET"),
|
||||||
|
("Zulu", "Zulu"),
|
||||||
|
("localtime", "localtime"),
|
||||||
|
],
|
||||||
|
default="UTC",
|
||||||
|
max_length=255,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
]
|
27
bookwyrm/migrations/0200_auto_20240327_1914.py
Normal file
27
bookwyrm/migrations/0200_auto_20240327_1914.py
Normal file
|
@ -0,0 +1,27 @@
|
||||||
|
# Generated by Django 3.2.25 on 2024-03-27 19:14
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0199_merge_20240326_1217"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.RemoveField(
|
||||||
|
model_name="addfiletotar",
|
||||||
|
name="childjob_ptr",
|
||||||
|
),
|
||||||
|
migrations.RemoveField(
|
||||||
|
model_name="addfiletotar",
|
||||||
|
name="parent_export_job",
|
||||||
|
),
|
||||||
|
migrations.DeleteModel(
|
||||||
|
name="AddBookToUserExportJob",
|
||||||
|
),
|
||||||
|
migrations.DeleteModel(
|
||||||
|
name="AddFileToTar",
|
||||||
|
),
|
||||||
|
]
|
|
@ -0,0 +1,39 @@
|
||||||
|
# Generated by Django 4.2.11 on 2024-04-01 21:09
|
||||||
|
|
||||||
|
import bookwyrm.models.fields
|
||||||
|
from django.db import migrations, models
|
||||||
|
from django.contrib.postgres.operations import CreateCollation
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0200_alter_user_preferred_timezone"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
CreateCollation(
|
||||||
|
"case_insensitive",
|
||||||
|
provider="icu",
|
||||||
|
locale="und-u-ks-level2",
|
||||||
|
deterministic=False,
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="hashtag",
|
||||||
|
name="name",
|
||||||
|
field=bookwyrm.models.fields.CharField(
|
||||||
|
db_collation="case_insensitive", max_length=256
|
||||||
|
),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name="user",
|
||||||
|
name="localname",
|
||||||
|
field=models.CharField(
|
||||||
|
db_collation="case_insensitive",
|
||||||
|
max_length=255,
|
||||||
|
null=True,
|
||||||
|
unique=True,
|
||||||
|
validators=[bookwyrm.models.fields.validate_localname],
|
||||||
|
),
|
||||||
|
),
|
||||||
|
]
|
13
bookwyrm/migrations/0204_merge_20240409_1042.py
Normal file
13
bookwyrm/migrations/0204_merge_20240409_1042.py
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
# Generated by Django 3.2.25 on 2024-04-09 10:42
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0197_mergedauthor_mergedbook"),
|
||||||
|
("bookwyrm", "0203_user_bookwyrm_us_is_acti_972dc4_idx"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = []
|
13
bookwyrm/migrations/0205_merge_20240410_2022.py
Normal file
13
bookwyrm/migrations/0205_merge_20240410_2022.py
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
# Generated by Django 4.2.11 on 2024-04-10 20:22
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0201_alter_hashtag_name_alter_user_localname"),
|
||||||
|
("bookwyrm", "0204_merge_20240409_1042"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = []
|
13
bookwyrm/migrations/0205_merge_20240413_0232.py
Normal file
13
bookwyrm/migrations/0205_merge_20240413_0232.py
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
# Generated by Django 3.2.25 on 2024-04-13 02:32
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0200_auto_20240327_1914"),
|
||||||
|
("bookwyrm", "0204_merge_20240409_1042"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = []
|
13
bookwyrm/migrations/0206_merge_20240415_1537.py
Normal file
13
bookwyrm/migrations/0206_merge_20240415_1537.py
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
# Generated by Django 4.2.11 on 2024-04-15 15:37
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("bookwyrm", "0205_merge_20240410_2022"),
|
||||||
|
("bookwyrm", "0205_merge_20240413_0232"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = []
|
|
@ -169,7 +169,7 @@ class ActivitypubMixin:
|
||||||
# filter users first by whether they're using the desired software
|
# filter users first by whether they're using the desired software
|
||||||
# this lets us send book updates only to other bw servers
|
# this lets us send book updates only to other bw servers
|
||||||
if software:
|
if software:
|
||||||
queryset = queryset.filter(bookwyrm_user=(software == "bookwyrm"))
|
queryset = queryset.filter(bookwyrm_user=software == "bookwyrm")
|
||||||
# if there's a user, we only want to send to the user's followers
|
# if there's a user, we only want to send to the user's followers
|
||||||
if user:
|
if user:
|
||||||
queryset = queryset.filter(following=user)
|
queryset = queryset.filter(following=user)
|
||||||
|
@ -206,14 +206,10 @@ class ObjectMixin(ActivitypubMixin):
|
||||||
created: Optional[bool] = None,
|
created: Optional[bool] = None,
|
||||||
software: Any = None,
|
software: Any = None,
|
||||||
priority: str = BROADCAST,
|
priority: str = BROADCAST,
|
||||||
|
broadcast: bool = True,
|
||||||
**kwargs: Any,
|
**kwargs: Any,
|
||||||
) -> None:
|
) -> None:
|
||||||
"""broadcast created/updated/deleted objects as appropriate"""
|
"""broadcast created/updated/deleted objects as appropriate"""
|
||||||
broadcast = kwargs.get("broadcast", True)
|
|
||||||
# this bonus kwarg would cause an error in the base save method
|
|
||||||
if "broadcast" in kwargs:
|
|
||||||
del kwargs["broadcast"]
|
|
||||||
|
|
||||||
created = created or not bool(self.id)
|
created = created or not bool(self.id)
|
||||||
# first off, we want to save normally no matter what
|
# first off, we want to save normally no matter what
|
||||||
super().save(*args, **kwargs)
|
super().save(*args, **kwargs)
|
||||||
|
|
|
@ -1,22 +1,25 @@
|
||||||
""" database schema for info about authors """
|
""" database schema for info about authors """
|
||||||
|
|
||||||
import re
|
import re
|
||||||
from typing import Tuple, Any
|
from typing import Any
|
||||||
|
|
||||||
from django.db import models
|
from django.db import models
|
||||||
from django.contrib.postgres.indexes import GinIndex
|
from django.contrib.postgres.indexes import GinIndex
|
||||||
import pgtrigger
|
import pgtrigger
|
||||||
|
|
||||||
from bookwyrm import activitypub
|
from bookwyrm import activitypub
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
from bookwyrm.utils.db import format_trigger
|
from bookwyrm.utils.db import format_trigger
|
||||||
|
|
||||||
from .book import BookDataModel
|
from .book import BookDataModel, MergedAuthor
|
||||||
from . import fields
|
from . import fields
|
||||||
|
|
||||||
|
|
||||||
class Author(BookDataModel):
|
class Author(BookDataModel):
|
||||||
"""basic biographic info"""
|
"""basic biographic info"""
|
||||||
|
|
||||||
|
merged_model = MergedAuthor
|
||||||
|
|
||||||
wikipedia_link = fields.CharField(
|
wikipedia_link = fields.CharField(
|
||||||
max_length=255, blank=True, null=True, deduplication_field=True
|
max_length=255, blank=True, null=True, deduplication_field=True
|
||||||
)
|
)
|
||||||
|
@ -42,12 +45,12 @@ class Author(BookDataModel):
|
||||||
)
|
)
|
||||||
bio = fields.HtmlField(null=True, blank=True)
|
bio = fields.HtmlField(null=True, blank=True)
|
||||||
|
|
||||||
def save(self, *args: Tuple[Any, ...], **kwargs: dict[str, Any]) -> None:
|
def save(self, *args: Any, **kwargs: Any) -> None:
|
||||||
"""normalize isni format"""
|
"""normalize isni format"""
|
||||||
if self.isni:
|
if self.isni is not None:
|
||||||
self.isni = re.sub(r"\s", "", self.isni)
|
self.isni = re.sub(r"\s", "", self.isni)
|
||||||
|
|
||||||
return super().save(*args, **kwargs)
|
super().save(*args, **kwargs)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def isni_link(self):
|
def isni_link(self):
|
||||||
|
@ -67,7 +70,7 @@ class Author(BookDataModel):
|
||||||
|
|
||||||
def get_remote_id(self):
|
def get_remote_id(self):
|
||||||
"""editions and works both use "book" instead of model_name"""
|
"""editions and works both use "book" instead of model_name"""
|
||||||
return f"https://{DOMAIN}/author/{self.id}"
|
return f"{BASE_URL}/author/{self.id}"
|
||||||
|
|
||||||
class Meta:
|
class Meta:
|
||||||
"""sets up indexes and triggers"""
|
"""sets up indexes and triggers"""
|
||||||
|
|
|
@ -10,7 +10,7 @@ from django.http import Http404
|
||||||
from django.utils.translation import gettext_lazy as _
|
from django.utils.translation import gettext_lazy as _
|
||||||
from django.utils.text import slugify
|
from django.utils.text import slugify
|
||||||
|
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
from .fields import RemoteIdField
|
from .fields import RemoteIdField
|
||||||
|
|
||||||
|
|
||||||
|
@ -38,7 +38,7 @@ class BookWyrmModel(models.Model):
|
||||||
|
|
||||||
def get_remote_id(self):
|
def get_remote_id(self):
|
||||||
"""generate the url that resolves to the local object, without a slug"""
|
"""generate the url that resolves to the local object, without a slug"""
|
||||||
base_path = f"https://{DOMAIN}"
|
base_path = BASE_URL
|
||||||
if hasattr(self, "user"):
|
if hasattr(self, "user"):
|
||||||
base_path = f"{base_path}{self.user.local_path}"
|
base_path = f"{base_path}{self.user.local_path}"
|
||||||
|
|
||||||
|
@ -53,7 +53,7 @@ class BookWyrmModel(models.Model):
|
||||||
@property
|
@property
|
||||||
def local_path(self):
|
def local_path(self):
|
||||||
"""how to link to this object in the local app, with a slug"""
|
"""how to link to this object in the local app, with a slug"""
|
||||||
local = self.get_remote_id().replace(f"https://{DOMAIN}", "")
|
local = self.get_remote_id().replace(BASE_URL, "")
|
||||||
|
|
||||||
name = None
|
name = None
|
||||||
if hasattr(self, "name_field"):
|
if hasattr(self, "name_field"):
|
||||||
|
|
|
@ -1,13 +1,15 @@
|
||||||
""" database schema for books and shelves """
|
""" database schema for books and shelves """
|
||||||
|
|
||||||
from itertools import chain
|
from itertools import chain
|
||||||
import re
|
import re
|
||||||
from typing import Any
|
from typing import Any, Dict, Optional, Iterable
|
||||||
|
from typing_extensions import Self
|
||||||
|
|
||||||
from django.contrib.postgres.search import SearchVectorField
|
from django.contrib.postgres.search import SearchVectorField
|
||||||
from django.contrib.postgres.indexes import GinIndex
|
from django.contrib.postgres.indexes import GinIndex
|
||||||
from django.core.cache import cache
|
from django.core.cache import cache
|
||||||
from django.db import models, transaction
|
from django.db import models, transaction
|
||||||
from django.db.models import Prefetch
|
from django.db.models import Prefetch, ManyToManyField
|
||||||
from django.dispatch import receiver
|
from django.dispatch import receiver
|
||||||
from django.utils.translation import gettext_lazy as _
|
from django.utils.translation import gettext_lazy as _
|
||||||
from model_utils import FieldTracker
|
from model_utils import FieldTracker
|
||||||
|
@ -19,13 +21,13 @@ from bookwyrm import activitypub
|
||||||
from bookwyrm.isbn.isbn import hyphenator_singleton as hyphenator
|
from bookwyrm.isbn.isbn import hyphenator_singleton as hyphenator
|
||||||
from bookwyrm.preview_images import generate_edition_preview_image_task
|
from bookwyrm.preview_images import generate_edition_preview_image_task
|
||||||
from bookwyrm.settings import (
|
from bookwyrm.settings import (
|
||||||
DOMAIN,
|
BASE_URL,
|
||||||
DEFAULT_LANGUAGE,
|
DEFAULT_LANGUAGE,
|
||||||
LANGUAGE_ARTICLES,
|
LANGUAGE_ARTICLES,
|
||||||
ENABLE_PREVIEW_IMAGES,
|
ENABLE_PREVIEW_IMAGES,
|
||||||
ENABLE_THUMBNAIL_GENERATION,
|
ENABLE_THUMBNAIL_GENERATION,
|
||||||
)
|
)
|
||||||
from bookwyrm.utils.db import format_trigger
|
from bookwyrm.utils.db import format_trigger, add_update_fields
|
||||||
|
|
||||||
from .activitypub_mixin import OrderedCollectionPageMixin, ObjectMixin
|
from .activitypub_mixin import OrderedCollectionPageMixin, ObjectMixin
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
|
@ -94,24 +96,134 @@ class BookDataModel(ObjectMixin, BookWyrmModel):
|
||||||
|
|
||||||
abstract = True
|
abstract = True
|
||||||
|
|
||||||
def save(self, *args: Any, **kwargs: Any) -> None:
|
def save(
|
||||||
|
self, *args: Any, update_fields: Optional[Iterable[str]] = None, **kwargs: Any
|
||||||
|
) -> None:
|
||||||
"""ensure that the remote_id is within this instance"""
|
"""ensure that the remote_id is within this instance"""
|
||||||
if self.id:
|
if self.id:
|
||||||
self.remote_id = self.get_remote_id()
|
self.remote_id = self.get_remote_id()
|
||||||
|
update_fields = add_update_fields(update_fields, "remote_id")
|
||||||
else:
|
else:
|
||||||
self.origin_id = self.remote_id
|
self.origin_id = self.remote_id
|
||||||
self.remote_id = None
|
self.remote_id = None
|
||||||
return super().save(*args, **kwargs)
|
update_fields = add_update_fields(update_fields, "origin_id", "remote_id")
|
||||||
|
|
||||||
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
# pylint: disable=arguments-differ
|
# pylint: disable=arguments-differ
|
||||||
def broadcast(self, activity, sender, software="bookwyrm", **kwargs):
|
def broadcast(self, activity, sender, software="bookwyrm", **kwargs):
|
||||||
"""only send book data updates to other bookwyrm instances"""
|
"""only send book data updates to other bookwyrm instances"""
|
||||||
super().broadcast(activity, sender, software=software, **kwargs)
|
super().broadcast(activity, sender, software=software, **kwargs)
|
||||||
|
|
||||||
|
def merge_into(self, canonical: Self, dry_run=False) -> Dict[str, Any]:
|
||||||
|
"""merge this entity into another entity"""
|
||||||
|
if canonical.id == self.id:
|
||||||
|
raise ValueError(f"Cannot merge {self} into itself")
|
||||||
|
|
||||||
|
absorbed_fields = canonical.absorb_data_from(self, dry_run=dry_run)
|
||||||
|
|
||||||
|
if dry_run:
|
||||||
|
return absorbed_fields
|
||||||
|
|
||||||
|
canonical.save()
|
||||||
|
|
||||||
|
self.merged_model.objects.create(deleted_id=self.id, merged_into=canonical)
|
||||||
|
|
||||||
|
# move related models to canonical
|
||||||
|
related_models = [
|
||||||
|
(r.remote_field.name, r.related_model) for r in self._meta.related_objects
|
||||||
|
]
|
||||||
|
# pylint: disable=protected-access
|
||||||
|
for related_field, related_model in related_models:
|
||||||
|
# Skip the ManyToMany fields that aren’t auto-created. These
|
||||||
|
# should have a corresponding OneToMany field in the model for
|
||||||
|
# the linking table anyway. If we update it through that model
|
||||||
|
# instead then we won’t lose the extra fields in the linking
|
||||||
|
# table.
|
||||||
|
# pylint: disable=protected-access
|
||||||
|
related_field_obj = related_model._meta.get_field(related_field)
|
||||||
|
if isinstance(related_field_obj, ManyToManyField):
|
||||||
|
through = related_field_obj.remote_field.through
|
||||||
|
if not through._meta.auto_created:
|
||||||
|
continue
|
||||||
|
related_objs = related_model.objects.filter(**{related_field: self})
|
||||||
|
for related_obj in related_objs:
|
||||||
|
try:
|
||||||
|
setattr(related_obj, related_field, canonical)
|
||||||
|
related_obj.save()
|
||||||
|
except TypeError:
|
||||||
|
getattr(related_obj, related_field).add(canonical)
|
||||||
|
getattr(related_obj, related_field).remove(self)
|
||||||
|
|
||||||
|
self.delete()
|
||||||
|
return absorbed_fields
|
||||||
|
|
||||||
|
def absorb_data_from(self, other: Self, dry_run=False) -> Dict[str, Any]:
|
||||||
|
"""fill empty fields with values from another entity"""
|
||||||
|
absorbed_fields = {}
|
||||||
|
for data_field in self._meta.get_fields():
|
||||||
|
if not hasattr(data_field, "activitypub_field"):
|
||||||
|
continue
|
||||||
|
canonical_value = getattr(self, data_field.name)
|
||||||
|
other_value = getattr(other, data_field.name)
|
||||||
|
if not other_value:
|
||||||
|
continue
|
||||||
|
if isinstance(data_field, fields.ArrayField):
|
||||||
|
if new_values := list(set(other_value) - set(canonical_value)):
|
||||||
|
# append at the end (in no particular order)
|
||||||
|
if not dry_run:
|
||||||
|
setattr(self, data_field.name, canonical_value + new_values)
|
||||||
|
absorbed_fields[data_field.name] = new_values
|
||||||
|
elif isinstance(data_field, fields.PartialDateField):
|
||||||
|
if (
|
||||||
|
(not canonical_value)
|
||||||
|
or (other_value.has_day and not canonical_value.has_day)
|
||||||
|
or (other_value.has_month and not canonical_value.has_month)
|
||||||
|
):
|
||||||
|
if not dry_run:
|
||||||
|
setattr(self, data_field.name, other_value)
|
||||||
|
absorbed_fields[data_field.name] = other_value
|
||||||
|
else:
|
||||||
|
if not canonical_value:
|
||||||
|
if not dry_run:
|
||||||
|
setattr(self, data_field.name, other_value)
|
||||||
|
absorbed_fields[data_field.name] = other_value
|
||||||
|
return absorbed_fields
|
||||||
|
|
||||||
|
|
||||||
|
class MergedBookDataModel(models.Model):
|
||||||
|
"""a BookDataModel instance that has been merged into another instance. kept
|
||||||
|
to be able to redirect old URLs"""
|
||||||
|
|
||||||
|
deleted_id = models.IntegerField(primary_key=True)
|
||||||
|
|
||||||
|
class Meta:
|
||||||
|
"""abstract just like BookDataModel"""
|
||||||
|
|
||||||
|
abstract = True
|
||||||
|
|
||||||
|
|
||||||
|
class MergedBook(MergedBookDataModel):
|
||||||
|
"""an Book that has been merged into another one"""
|
||||||
|
|
||||||
|
merged_into = models.ForeignKey(
|
||||||
|
"Book", on_delete=models.PROTECT, related_name="absorbed"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class MergedAuthor(MergedBookDataModel):
|
||||||
|
"""an Author that has been merged into another one"""
|
||||||
|
|
||||||
|
merged_into = models.ForeignKey(
|
||||||
|
"Author", on_delete=models.PROTECT, related_name="absorbed"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class Book(BookDataModel):
|
class Book(BookDataModel):
|
||||||
"""a generic book, which can mean either an edition or a work"""
|
"""a generic book, which can mean either an edition or a work"""
|
||||||
|
|
||||||
|
merged_model = MergedBook
|
||||||
|
|
||||||
connector = models.ForeignKey("Connector", on_delete=models.PROTECT, null=True)
|
connector = models.ForeignKey("Connector", on_delete=models.PROTECT, null=True)
|
||||||
|
|
||||||
# book/work metadata
|
# book/work metadata
|
||||||
|
@ -192,9 +304,13 @@ class Book(BookDataModel):
|
||||||
"""properties of this edition, as a string"""
|
"""properties of this edition, as a string"""
|
||||||
items = [
|
items = [
|
||||||
self.physical_format if hasattr(self, "physical_format") else None,
|
self.physical_format if hasattr(self, "physical_format") else None,
|
||||||
f"{self.languages[0]} language"
|
(
|
||||||
if self.languages and self.languages[0] and self.languages[0] != "English"
|
f"{self.languages[0]} language"
|
||||||
else None,
|
if self.languages
|
||||||
|
and self.languages[0]
|
||||||
|
and self.languages[0] != "English"
|
||||||
|
else None
|
||||||
|
),
|
||||||
str(self.published_date.year) if self.published_date else None,
|
str(self.published_date.year) if self.published_date else None,
|
||||||
", ".join(self.publishers) if hasattr(self, "publishers") else None,
|
", ".join(self.publishers) if hasattr(self, "publishers") else None,
|
||||||
]
|
]
|
||||||
|
@ -212,11 +328,11 @@ class Book(BookDataModel):
|
||||||
if not isinstance(self, (Edition, Work)):
|
if not isinstance(self, (Edition, Work)):
|
||||||
raise ValueError("Books should be added as Editions or Works")
|
raise ValueError("Books should be added as Editions or Works")
|
||||||
|
|
||||||
return super().save(*args, **kwargs)
|
super().save(*args, **kwargs)
|
||||||
|
|
||||||
def get_remote_id(self):
|
def get_remote_id(self):
|
||||||
"""editions and works both use "book" instead of model_name"""
|
"""editions and works both use "book" instead of model_name"""
|
||||||
return f"https://{DOMAIN}/book/{self.id}"
|
return f"{BASE_URL}/book/{self.id}"
|
||||||
|
|
||||||
def guess_sort_title(self):
|
def guess_sort_title(self):
|
||||||
"""Get a best-guess sort title for the current book"""
|
"""Get a best-guess sort title for the current book"""
|
||||||
|
@ -289,10 +405,11 @@ class Work(OrderedCollectionPageMixin, Book):
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, **kwargs):
|
||||||
"""set some fields on the edition object"""
|
"""set some fields on the edition object"""
|
||||||
|
super().save(*args, **kwargs)
|
||||||
|
|
||||||
# set rank
|
# set rank
|
||||||
for edition in self.editions.all():
|
for edition in self.editions.all():
|
||||||
edition.save()
|
edition.save()
|
||||||
return super().save(*args, **kwargs)
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def default_edition(self):
|
def default_edition(self):
|
||||||
|
@ -398,33 +515,48 @@ class Edition(Book):
|
||||||
# max rank is 9
|
# max rank is 9
|
||||||
return rank
|
return rank
|
||||||
|
|
||||||
def save(self, *args: Any, **kwargs: Any) -> None:
|
def save(
|
||||||
|
self, *args: Any, update_fields: Optional[Iterable[str]] = None, **kwargs: Any
|
||||||
|
) -> None:
|
||||||
"""set some fields on the edition object"""
|
"""set some fields on the edition object"""
|
||||||
# calculate isbn 10/13
|
# calculate isbn 10/13
|
||||||
if self.isbn_13 and self.isbn_13[:3] == "978" and not self.isbn_10:
|
if (
|
||||||
|
self.isbn_10 is None
|
||||||
|
and self.isbn_13 is not None
|
||||||
|
and self.isbn_13[:3] == "978"
|
||||||
|
):
|
||||||
self.isbn_10 = isbn_13_to_10(self.isbn_13)
|
self.isbn_10 = isbn_13_to_10(self.isbn_13)
|
||||||
if self.isbn_10 and not self.isbn_13:
|
update_fields = add_update_fields(update_fields, "isbn_10")
|
||||||
|
if self.isbn_13 is None and self.isbn_10 is not None:
|
||||||
self.isbn_13 = isbn_10_to_13(self.isbn_10)
|
self.isbn_13 = isbn_10_to_13(self.isbn_10)
|
||||||
|
update_fields = add_update_fields(update_fields, "isbn_13")
|
||||||
|
|
||||||
# normalize isbn format
|
# normalize isbn format
|
||||||
if self.isbn_10:
|
if self.isbn_10 is not None:
|
||||||
self.isbn_10 = normalize_isbn(self.isbn_10)
|
self.isbn_10 = normalize_isbn(self.isbn_10)
|
||||||
if self.isbn_13:
|
if self.isbn_13 is not None:
|
||||||
self.isbn_13 = normalize_isbn(self.isbn_13)
|
self.isbn_13 = normalize_isbn(self.isbn_13)
|
||||||
|
|
||||||
# set rank
|
# set rank
|
||||||
self.edition_rank = self.get_rank()
|
if (new := self.get_rank()) != self.edition_rank:
|
||||||
|
self.edition_rank = new
|
||||||
# clear author cache
|
update_fields = add_update_fields(update_fields, "edition_rank")
|
||||||
if self.id:
|
|
||||||
for author_id in self.authors.values_list("id", flat=True):
|
|
||||||
cache.delete(f"author-books-{author_id}")
|
|
||||||
|
|
||||||
# Create sort title by removing articles from title
|
# Create sort title by removing articles from title
|
||||||
if self.sort_title in [None, ""]:
|
if self.sort_title in [None, ""]:
|
||||||
self.sort_title = self.guess_sort_title()
|
self.sort_title = self.guess_sort_title()
|
||||||
|
update_fields = add_update_fields(update_fields, "sort_title")
|
||||||
|
|
||||||
return super().save(*args, **kwargs)
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
|
# clear author cache
|
||||||
|
if self.id:
|
||||||
|
cache.delete_many(
|
||||||
|
[
|
||||||
|
f"author-books-{author_id}"
|
||||||
|
for author_id in self.authors.values_list("id", flat=True)
|
||||||
|
]
|
||||||
|
)
|
||||||
|
|
||||||
@transaction.atomic
|
@transaction.atomic
|
||||||
def repair(self):
|
def repair(self):
|
||||||
|
|
|
@ -1,213 +1,317 @@
|
||||||
"""Export user account to tar.gz file for import into another Bookwyrm instance"""
|
"""Export user account to tar.gz file for import into another Bookwyrm instance"""
|
||||||
|
|
||||||
import dataclasses
|
|
||||||
import logging
|
import logging
|
||||||
from uuid import uuid4
|
import os
|
||||||
|
|
||||||
from django.db.models import FileField
|
from boto3.session import Session as BotoSession
|
||||||
|
from s3_tar import S3Tar
|
||||||
|
|
||||||
|
from django.db.models import BooleanField, FileField, JSONField
|
||||||
from django.db.models import Q
|
from django.db.models import Q
|
||||||
from django.core.serializers.json import DjangoJSONEncoder
|
from django.core.serializers.json import DjangoJSONEncoder
|
||||||
from django.core.files.base import ContentFile
|
from django.core.files.base import ContentFile
|
||||||
|
from django.core.files.storage import storages
|
||||||
|
|
||||||
from bookwyrm.models import AnnualGoal, ReadThrough, ShelfBook, List, ListItem
|
from bookwyrm import settings
|
||||||
|
|
||||||
|
from bookwyrm.models import AnnualGoal, ReadThrough, ShelfBook, ListItem
|
||||||
from bookwyrm.models import Review, Comment, Quotation
|
from bookwyrm.models import Review, Comment, Quotation
|
||||||
from bookwyrm.models import Edition
|
from bookwyrm.models import Edition
|
||||||
from bookwyrm.models import UserFollows, User, UserBlocks
|
from bookwyrm.models import UserFollows, User, UserBlocks
|
||||||
from bookwyrm.models.job import ParentJob, ParentTask
|
from bookwyrm.models.job import ParentJob
|
||||||
from bookwyrm.tasks import app, IMPORTS
|
from bookwyrm.tasks import app, IMPORTS
|
||||||
from bookwyrm.utils.tar import BookwyrmTarFile
|
from bookwyrm.utils.tar import BookwyrmTarFile
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class BookwyrmAwsSession(BotoSession):
|
||||||
|
"""a boto session that always uses settings.AWS_S3_ENDPOINT_URL"""
|
||||||
|
|
||||||
|
def client(self, *args, **kwargs): # pylint: disable=arguments-differ
|
||||||
|
kwargs["endpoint_url"] = settings.AWS_S3_ENDPOINT_URL
|
||||||
|
return super().client("s3", *args, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
def select_exports_storage():
|
||||||
|
"""callable to allow for dependency on runtime configuration"""
|
||||||
|
return storages["exports"]
|
||||||
|
|
||||||
|
|
||||||
class BookwyrmExportJob(ParentJob):
|
class BookwyrmExportJob(ParentJob):
|
||||||
"""entry for a specific request to export a bookwyrm user"""
|
"""entry for a specific request to export a bookwyrm user"""
|
||||||
|
|
||||||
export_data = FileField(null=True)
|
export_data = FileField(null=True, storage=select_exports_storage)
|
||||||
|
export_json = JSONField(null=True, encoder=DjangoJSONEncoder)
|
||||||
|
json_completed = BooleanField(default=False)
|
||||||
|
|
||||||
def start_job(self):
|
def start_job(self):
|
||||||
"""Start the job"""
|
"""schedule the first task"""
|
||||||
start_export_task.delay(job_id=self.id, no_children=True)
|
|
||||||
|
|
||||||
return self
|
task = create_export_json_task.delay(job_id=self.id)
|
||||||
|
self.task_id = task.id
|
||||||
|
self.save(update_fields=["task_id"])
|
||||||
|
|
||||||
|
|
||||||
@app.task(queue=IMPORTS, base=ParentTask)
|
@app.task(queue=IMPORTS)
|
||||||
def start_export_task(**kwargs):
|
def create_export_json_task(job_id):
|
||||||
"""trigger the child tasks for each row"""
|
"""create the JSON data for the export"""
|
||||||
job = BookwyrmExportJob.objects.get(id=kwargs["job_id"])
|
|
||||||
|
job = BookwyrmExportJob.objects.get(id=job_id)
|
||||||
|
|
||||||
# don't start the job if it was stopped from the UI
|
# don't start the job if it was stopped from the UI
|
||||||
if job.complete:
|
if job.complete:
|
||||||
return
|
return
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# This is where ChildJobs get made
|
job.set_status("active")
|
||||||
job.export_data = ContentFile(b"", str(uuid4()))
|
|
||||||
json_data = json_export(job.user)
|
# generate JSON structure
|
||||||
tar_export(json_data, job.user, job.export_data)
|
job.export_json = export_json(job.user)
|
||||||
job.save(update_fields=["export_data"])
|
job.save(update_fields=["export_json"])
|
||||||
|
|
||||||
|
# create archive in separate task
|
||||||
|
create_archive_task.delay(job_id=job.id)
|
||||||
except Exception as err: # pylint: disable=broad-except
|
except Exception as err: # pylint: disable=broad-except
|
||||||
logger.exception("User Export Job %s Failed with error: %s", job.id, err)
|
logger.exception(
|
||||||
|
"create_export_json_task for %s failed with error: %s", job, err
|
||||||
|
)
|
||||||
job.set_status("failed")
|
job.set_status("failed")
|
||||||
|
|
||||||
job.set_status("complete")
|
|
||||||
|
def archive_file_location(file, directory="") -> str:
|
||||||
|
"""get the relative location of a file inside the archive"""
|
||||||
|
return os.path.join(directory, file.name)
|
||||||
|
|
||||||
|
|
||||||
def tar_export(json_data: str, user, file):
|
def add_file_to_s3_tar(s3_tar: S3Tar, storage, file, directory=""):
|
||||||
"""wrap the export information in a tar file"""
|
"""
|
||||||
file.open("wb")
|
add file to S3Tar inside directory, keeping any directories under its
|
||||||
with BookwyrmTarFile.open(mode="w:gz", fileobj=file) as tar:
|
storage location
|
||||||
tar.write_bytes(json_data.encode("utf-8"))
|
"""
|
||||||
|
s3_tar.add_file(
|
||||||
|
os.path.join(storage.location, file.name),
|
||||||
|
folder=os.path.dirname(archive_file_location(file, directory=directory)),
|
||||||
|
)
|
||||||
|
|
||||||
# Add avatar image if present
|
|
||||||
if getattr(user, "avatar", False):
|
|
||||||
tar.add_image(user.avatar, filename="avatar")
|
|
||||||
|
|
||||||
|
@app.task(queue=IMPORTS)
|
||||||
|
def create_archive_task(job_id):
|
||||||
|
"""create the archive containing the JSON file and additional files"""
|
||||||
|
|
||||||
|
job = BookwyrmExportJob.objects.get(id=job_id)
|
||||||
|
|
||||||
|
# don't start the job if it was stopped from the UI
|
||||||
|
if job.complete:
|
||||||
|
return
|
||||||
|
|
||||||
|
try:
|
||||||
|
export_task_id = str(job.task_id)
|
||||||
|
archive_filename = f"{export_task_id}.tar.gz"
|
||||||
|
export_json_bytes = DjangoJSONEncoder().encode(job.export_json).encode("utf-8")
|
||||||
|
|
||||||
|
user = job.user
|
||||||
editions = get_books_for_user(user)
|
editions = get_books_for_user(user)
|
||||||
for book in editions:
|
|
||||||
if getattr(book, "cover", False):
|
|
||||||
tar.add_image(book.cover)
|
|
||||||
|
|
||||||
file.close()
|
if settings.USE_S3:
|
||||||
|
# Storage for writing temporary files
|
||||||
|
exports_storage = storages["exports"]
|
||||||
|
|
||||||
|
# Handle for creating the final archive
|
||||||
|
s3_tar = S3Tar(
|
||||||
|
exports_storage.bucket_name,
|
||||||
|
os.path.join(exports_storage.location, archive_filename),
|
||||||
|
session=BookwyrmAwsSession(),
|
||||||
|
)
|
||||||
|
|
||||||
|
# Save JSON file to a temporary location
|
||||||
|
export_json_tmp_file = os.path.join(export_task_id, "archive.json")
|
||||||
|
exports_storage.save(
|
||||||
|
export_json_tmp_file,
|
||||||
|
ContentFile(export_json_bytes),
|
||||||
|
)
|
||||||
|
s3_tar.add_file(
|
||||||
|
os.path.join(exports_storage.location, export_json_tmp_file)
|
||||||
|
)
|
||||||
|
|
||||||
|
# Add images to TAR
|
||||||
|
images_storage = storages["default"]
|
||||||
|
|
||||||
|
if user.avatar:
|
||||||
|
add_file_to_s3_tar(s3_tar, images_storage, user.avatar)
|
||||||
|
|
||||||
|
for edition in editions:
|
||||||
|
if edition.cover:
|
||||||
|
add_file_to_s3_tar(
|
||||||
|
s3_tar, images_storage, edition.cover, directory="images"
|
||||||
|
)
|
||||||
|
|
||||||
|
# Create archive and store file name
|
||||||
|
s3_tar.tar()
|
||||||
|
job.export_data = archive_filename
|
||||||
|
job.save(update_fields=["export_data"])
|
||||||
|
|
||||||
|
# Delete temporary files
|
||||||
|
exports_storage.delete(export_json_tmp_file)
|
||||||
|
|
||||||
|
else:
|
||||||
|
job.export_data = archive_filename
|
||||||
|
with job.export_data.open("wb") as tar_file:
|
||||||
|
with BookwyrmTarFile.open(mode="w:gz", fileobj=tar_file) as tar:
|
||||||
|
# save json file
|
||||||
|
tar.write_bytes(export_json_bytes)
|
||||||
|
|
||||||
|
# Add avatar image if present
|
||||||
|
if user.avatar:
|
||||||
|
tar.add_image(user.avatar)
|
||||||
|
|
||||||
|
for edition in editions:
|
||||||
|
if edition.cover:
|
||||||
|
tar.add_image(edition.cover, directory="images")
|
||||||
|
job.save(update_fields=["export_data"])
|
||||||
|
|
||||||
|
job.set_status("completed")
|
||||||
|
|
||||||
|
except Exception as err: # pylint: disable=broad-except
|
||||||
|
logger.exception("create_archive_task for %s failed with error: %s", job, err)
|
||||||
|
job.set_status("failed")
|
||||||
|
|
||||||
|
|
||||||
def json_export(
|
def export_json(user: User):
|
||||||
user,
|
"""create export JSON"""
|
||||||
): # pylint: disable=too-many-locals, too-many-statements, too-many-branches
|
data = export_user(user) # in the root of the JSON structure
|
||||||
"""Generate an export for a user"""
|
data["settings"] = export_settings(user)
|
||||||
|
data["goals"] = export_goals(user)
|
||||||
|
data["books"] = export_books(user)
|
||||||
|
data["saved_lists"] = export_saved_lists(user)
|
||||||
|
data["follows"] = export_follows(user)
|
||||||
|
data["blocks"] = export_blocks(user)
|
||||||
|
return data
|
||||||
|
|
||||||
# User as AP object
|
|
||||||
exported_user = user.to_activity()
|
def export_user(user: User):
|
||||||
# I don't love this but it prevents a JSON encoding error
|
"""export user data"""
|
||||||
# when there is no user image
|
data = user.to_activity()
|
||||||
if exported_user.get("icon") in (None, dataclasses.MISSING):
|
if user.avatar:
|
||||||
exported_user["icon"] = {}
|
data["icon"]["url"] = archive_file_location(user.avatar)
|
||||||
else:
|
else:
|
||||||
# change the URL to be relative to the JSON file
|
data["icon"] = {}
|
||||||
file_type = exported_user["icon"]["url"].rsplit(".", maxsplit=1)[-1]
|
return data
|
||||||
filename = f"avatar.{file_type}"
|
|
||||||
exported_user["icon"]["url"] = filename
|
|
||||||
|
|
||||||
# Additional settings - can't be serialized as AP
|
|
||||||
|
def export_settings(user: User):
|
||||||
|
"""Additional settings - can't be serialized as AP"""
|
||||||
vals = [
|
vals = [
|
||||||
"show_goal",
|
"show_goal",
|
||||||
"preferred_timezone",
|
"preferred_timezone",
|
||||||
"default_post_privacy",
|
"default_post_privacy",
|
||||||
"show_suggested_users",
|
"show_suggested_users",
|
||||||
]
|
]
|
||||||
exported_user["settings"] = {}
|
return {k: getattr(user, k) for k in vals}
|
||||||
for k in vals:
|
|
||||||
exported_user["settings"][k] = getattr(user, k)
|
|
||||||
|
|
||||||
# Reading goals - can't be serialized as AP
|
|
||||||
reading_goals = AnnualGoal.objects.filter(user=user).distinct()
|
|
||||||
exported_user["goals"] = []
|
|
||||||
for goal in reading_goals:
|
|
||||||
exported_user["goals"].append(
|
|
||||||
{"goal": goal.goal, "year": goal.year, "privacy": goal.privacy}
|
|
||||||
)
|
|
||||||
|
|
||||||
# Reading history - can't be serialized as AP
|
def export_saved_lists(user: User):
|
||||||
readthroughs = ReadThrough.objects.filter(user=user).distinct().values()
|
"""add user saved lists to export JSON"""
|
||||||
readthroughs = list(readthroughs)
|
return [l.remote_id for l in user.saved_lists.all()]
|
||||||
|
|
||||||
# Books
|
|
||||||
editions = get_books_for_user(user)
|
|
||||||
exported_user["books"] = []
|
|
||||||
|
|
||||||
for edition in editions:
|
def export_follows(user: User):
|
||||||
book = {}
|
"""add user follows to export JSON"""
|
||||||
book["work"] = edition.parent_work.to_activity()
|
|
||||||
book["edition"] = edition.to_activity()
|
|
||||||
|
|
||||||
if book["edition"].get("cover"):
|
|
||||||
# change the URL to be relative to the JSON file
|
|
||||||
filename = book["edition"]["cover"]["url"].rsplit("/", maxsplit=1)[-1]
|
|
||||||
book["edition"]["cover"]["url"] = f"covers/{filename}"
|
|
||||||
|
|
||||||
# authors
|
|
||||||
book["authors"] = []
|
|
||||||
for author in edition.authors.all():
|
|
||||||
book["authors"].append(author.to_activity())
|
|
||||||
|
|
||||||
# Shelves this book is on
|
|
||||||
# Every ShelfItem is this book so we don't other serializing
|
|
||||||
book["shelves"] = []
|
|
||||||
shelf_books = (
|
|
||||||
ShelfBook.objects.select_related("shelf")
|
|
||||||
.filter(user=user, book=edition)
|
|
||||||
.distinct()
|
|
||||||
)
|
|
||||||
|
|
||||||
for shelfbook in shelf_books:
|
|
||||||
book["shelves"].append(shelfbook.shelf.to_activity())
|
|
||||||
|
|
||||||
# Lists and ListItems
|
|
||||||
# ListItems include "notes" and "approved" so we need them
|
|
||||||
# even though we know it's this book
|
|
||||||
book["lists"] = []
|
|
||||||
list_items = ListItem.objects.filter(book=edition, user=user).distinct()
|
|
||||||
|
|
||||||
for item in list_items:
|
|
||||||
list_info = item.book_list.to_activity()
|
|
||||||
list_info[
|
|
||||||
"privacy"
|
|
||||||
] = item.book_list.privacy # this isn't serialized so we add it
|
|
||||||
list_info["list_item"] = item.to_activity()
|
|
||||||
book["lists"].append(list_info)
|
|
||||||
|
|
||||||
# Statuses
|
|
||||||
# Can't use select_subclasses here because
|
|
||||||
# we need to filter on the "book" value,
|
|
||||||
# which is not available on an ordinary Status
|
|
||||||
for status in ["comments", "quotations", "reviews"]:
|
|
||||||
book[status] = []
|
|
||||||
|
|
||||||
comments = Comment.objects.filter(user=user, book=edition).all()
|
|
||||||
for status in comments:
|
|
||||||
obj = status.to_activity()
|
|
||||||
obj["progress"] = status.progress
|
|
||||||
obj["progress_mode"] = status.progress_mode
|
|
||||||
book["comments"].append(obj)
|
|
||||||
|
|
||||||
quotes = Quotation.objects.filter(user=user, book=edition).all()
|
|
||||||
for status in quotes:
|
|
||||||
obj = status.to_activity()
|
|
||||||
obj["position"] = status.position
|
|
||||||
obj["endposition"] = status.endposition
|
|
||||||
obj["position_mode"] = status.position_mode
|
|
||||||
book["quotations"].append(obj)
|
|
||||||
|
|
||||||
reviews = Review.objects.filter(user=user, book=edition).all()
|
|
||||||
for status in reviews:
|
|
||||||
obj = status.to_activity()
|
|
||||||
book["reviews"].append(obj)
|
|
||||||
|
|
||||||
# readthroughs can't be serialized to activity
|
|
||||||
book_readthroughs = (
|
|
||||||
ReadThrough.objects.filter(user=user, book=edition).distinct().values()
|
|
||||||
)
|
|
||||||
book["readthroughs"] = list(book_readthroughs)
|
|
||||||
|
|
||||||
# append everything
|
|
||||||
exported_user["books"].append(book)
|
|
||||||
|
|
||||||
# saved book lists - just the remote id
|
|
||||||
saved_lists = List.objects.filter(id__in=user.saved_lists.all()).distinct()
|
|
||||||
exported_user["saved_lists"] = [l.remote_id for l in saved_lists]
|
|
||||||
|
|
||||||
# follows - just the remote id
|
|
||||||
follows = UserFollows.objects.filter(user_subject=user).distinct()
|
follows = UserFollows.objects.filter(user_subject=user).distinct()
|
||||||
following = User.objects.filter(userfollows_user_object__in=follows).distinct()
|
following = User.objects.filter(userfollows_user_object__in=follows).distinct()
|
||||||
exported_user["follows"] = [f.remote_id for f in following]
|
return [f.remote_id for f in following]
|
||||||
|
|
||||||
# blocks - just the remote id
|
|
||||||
|
def export_blocks(user: User):
|
||||||
|
"""add user blocks to export JSON"""
|
||||||
blocks = UserBlocks.objects.filter(user_subject=user).distinct()
|
blocks = UserBlocks.objects.filter(user_subject=user).distinct()
|
||||||
blocking = User.objects.filter(userblocks_user_object__in=blocks).distinct()
|
blocking = User.objects.filter(userblocks_user_object__in=blocks).distinct()
|
||||||
|
return [b.remote_id for b in blocking]
|
||||||
|
|
||||||
exported_user["blocks"] = [b.remote_id for b in blocking]
|
|
||||||
|
|
||||||
return DjangoJSONEncoder().encode(exported_user)
|
def export_goals(user: User):
|
||||||
|
"""add user reading goals to export JSON"""
|
||||||
|
reading_goals = AnnualGoal.objects.filter(user=user).distinct()
|
||||||
|
return [
|
||||||
|
{"goal": goal.goal, "year": goal.year, "privacy": goal.privacy}
|
||||||
|
for goal in reading_goals
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def export_books(user: User):
|
||||||
|
"""add books to export JSON"""
|
||||||
|
editions = get_books_for_user(user)
|
||||||
|
return [export_book(user, edition) for edition in editions]
|
||||||
|
|
||||||
|
|
||||||
|
def export_book(user: User, edition: Edition):
|
||||||
|
"""add book to export JSON"""
|
||||||
|
data = {}
|
||||||
|
data["work"] = edition.parent_work.to_activity()
|
||||||
|
data["edition"] = edition.to_activity()
|
||||||
|
|
||||||
|
if edition.cover:
|
||||||
|
data["edition"]["cover"]["url"] = archive_file_location(
|
||||||
|
edition.cover, directory="images"
|
||||||
|
)
|
||||||
|
|
||||||
|
# authors
|
||||||
|
data["authors"] = [author.to_activity() for author in edition.authors.all()]
|
||||||
|
|
||||||
|
# Shelves this book is on
|
||||||
|
# Every ShelfItem is this book so we don't other serializing
|
||||||
|
shelf_books = (
|
||||||
|
ShelfBook.objects.select_related("shelf")
|
||||||
|
.filter(user=user, book=edition)
|
||||||
|
.distinct()
|
||||||
|
)
|
||||||
|
data["shelves"] = [shelfbook.shelf.to_activity() for shelfbook in shelf_books]
|
||||||
|
|
||||||
|
# Lists and ListItems
|
||||||
|
# ListItems include "notes" and "approved" so we need them
|
||||||
|
# even though we know it's this book
|
||||||
|
list_items = ListItem.objects.filter(book=edition, user=user).distinct()
|
||||||
|
|
||||||
|
data["lists"] = []
|
||||||
|
for item in list_items:
|
||||||
|
list_info = item.book_list.to_activity()
|
||||||
|
list_info[
|
||||||
|
"privacy"
|
||||||
|
] = item.book_list.privacy # this isn't serialized so we add it
|
||||||
|
list_info["list_item"] = item.to_activity()
|
||||||
|
data["lists"].append(list_info)
|
||||||
|
|
||||||
|
# Statuses
|
||||||
|
# Can't use select_subclasses here because
|
||||||
|
# we need to filter on the "book" value,
|
||||||
|
# which is not available on an ordinary Status
|
||||||
|
for status in ["comments", "quotations", "reviews"]:
|
||||||
|
data[status] = []
|
||||||
|
|
||||||
|
comments = Comment.objects.filter(user=user, book=edition).all()
|
||||||
|
for status in comments:
|
||||||
|
obj = status.to_activity()
|
||||||
|
obj["progress"] = status.progress
|
||||||
|
obj["progress_mode"] = status.progress_mode
|
||||||
|
data["comments"].append(obj)
|
||||||
|
|
||||||
|
quotes = Quotation.objects.filter(user=user, book=edition).all()
|
||||||
|
for status in quotes:
|
||||||
|
obj = status.to_activity()
|
||||||
|
obj["position"] = status.position
|
||||||
|
obj["endposition"] = status.endposition
|
||||||
|
obj["position_mode"] = status.position_mode
|
||||||
|
data["quotations"].append(obj)
|
||||||
|
|
||||||
|
reviews = Review.objects.filter(user=user, book=edition).all()
|
||||||
|
data["reviews"] = [status.to_activity() for status in reviews]
|
||||||
|
|
||||||
|
# readthroughs can't be serialized to activity
|
||||||
|
book_readthroughs = (
|
||||||
|
ReadThrough.objects.filter(user=user, book=edition).distinct().values()
|
||||||
|
)
|
||||||
|
data["readthroughs"] = list(book_readthroughs)
|
||||||
|
return data
|
||||||
|
|
||||||
|
|
||||||
def get_books_for_user(user):
|
def get_books_for_user(user):
|
||||||
|
|
|
@ -42,20 +42,23 @@ def start_import_task(**kwargs):
|
||||||
try:
|
try:
|
||||||
archive_file.open("rb")
|
archive_file.open("rb")
|
||||||
with BookwyrmTarFile.open(mode="r:gz", fileobj=archive_file) as tar:
|
with BookwyrmTarFile.open(mode="r:gz", fileobj=archive_file) as tar:
|
||||||
job.import_data = json.loads(tar.read("archive.json").decode("utf-8"))
|
json_filename = next(
|
||||||
|
filter(lambda n: n.startswith("archive"), tar.getnames())
|
||||||
|
)
|
||||||
|
job.import_data = json.loads(tar.read(json_filename).decode("utf-8"))
|
||||||
|
|
||||||
if "include_user_profile" in job.required:
|
if "include_user_profile" in job.required:
|
||||||
update_user_profile(job.user, tar, job.import_data)
|
update_user_profile(job.user, tar, job.import_data)
|
||||||
if "include_user_settings" in job.required:
|
if "include_user_settings" in job.required:
|
||||||
update_user_settings(job.user, job.import_data)
|
update_user_settings(job.user, job.import_data)
|
||||||
if "include_goals" in job.required:
|
if "include_goals" in job.required:
|
||||||
update_goals(job.user, job.import_data.get("goals"))
|
update_goals(job.user, job.import_data.get("goals", []))
|
||||||
if "include_saved_lists" in job.required:
|
if "include_saved_lists" in job.required:
|
||||||
upsert_saved_lists(job.user, job.import_data.get("saved_lists"))
|
upsert_saved_lists(job.user, job.import_data.get("saved_lists", []))
|
||||||
if "include_follows" in job.required:
|
if "include_follows" in job.required:
|
||||||
upsert_follows(job.user, job.import_data.get("follows"))
|
upsert_follows(job.user, job.import_data.get("follows", []))
|
||||||
if "include_blocks" in job.required:
|
if "include_blocks" in job.required:
|
||||||
upsert_user_blocks(job.user, job.import_data.get("blocks"))
|
upsert_user_blocks(job.user, job.import_data.get("blocks", []))
|
||||||
|
|
||||||
process_books(job, tar)
|
process_books(job, tar)
|
||||||
|
|
||||||
|
@ -212,7 +215,7 @@ def upsert_statuses(user, cls, data, book_remote_id):
|
||||||
instance.save() # save and broadcast
|
instance.save() # save and broadcast
|
||||||
|
|
||||||
else:
|
else:
|
||||||
logger.info("User does not have permission to import statuses")
|
logger.warning("User does not have permission to import statuses")
|
||||||
|
|
||||||
|
|
||||||
def upsert_lists(user, lists, book_id):
|
def upsert_lists(user, lists, book_id):
|
||||||
|
|
|
@ -11,7 +11,7 @@ ConnectorFiles = models.TextChoices("ConnectorFiles", CONNECTORS)
|
||||||
class Connector(BookWyrmModel):
|
class Connector(BookWyrmModel):
|
||||||
"""book data source connectors"""
|
"""book data source connectors"""
|
||||||
|
|
||||||
identifier = models.CharField(max_length=255, unique=True)
|
identifier = models.CharField(max_length=255, unique=True) # domain
|
||||||
priority = models.IntegerField(default=2)
|
priority = models.IntegerField(default=2)
|
||||||
name = models.CharField(max_length=255, null=True, blank=True)
|
name = models.CharField(max_length=255, null=True, blank=True)
|
||||||
connector_file = models.CharField(max_length=255, choices=ConnectorFiles.choices)
|
connector_file = models.CharField(max_length=255, choices=ConnectorFiles.choices)
|
||||||
|
|
|
@ -16,7 +16,7 @@ FederationStatus = [
|
||||||
class FederatedServer(BookWyrmModel):
|
class FederatedServer(BookWyrmModel):
|
||||||
"""store which servers we federate with"""
|
"""store which servers we federate with"""
|
||||||
|
|
||||||
server_name = models.CharField(max_length=255, unique=True)
|
server_name = models.CharField(max_length=255, unique=True) # domain
|
||||||
status = models.CharField(
|
status = models.CharField(
|
||||||
max_length=255, default="federated", choices=FederationStatus
|
max_length=255, default="federated", choices=FederationStatus
|
||||||
)
|
)
|
||||||
|
@ -64,5 +64,4 @@ class FederatedServer(BookWyrmModel):
|
||||||
def is_blocked(cls, url: str) -> bool:
|
def is_blocked(cls, url: str) -> bool:
|
||||||
"""look up if a domain is blocked"""
|
"""look up if a domain is blocked"""
|
||||||
url = urlparse(url)
|
url = urlparse(url)
|
||||||
domain = url.netloc
|
return cls.objects.filter(server_name=url.hostname, status="blocked").exists()
|
||||||
return cls.objects.filter(server_name=domain, status="blocked").exists()
|
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
""" do book related things with other users """
|
""" do book related things with other users """
|
||||||
from django.db import models, IntegrityError, transaction
|
from django.db import models, IntegrityError, transaction
|
||||||
from django.db.models import Q
|
from django.db.models import Q
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
from . import fields
|
from . import fields
|
||||||
from .relationship import UserBlocks
|
from .relationship import UserBlocks
|
||||||
|
@ -17,7 +17,7 @@ class Group(BookWyrmModel):
|
||||||
|
|
||||||
def get_remote_id(self):
|
def get_remote_id(self):
|
||||||
"""don't want the user to be in there in this case"""
|
"""don't want the user to be in there in this case"""
|
||||||
return f"https://{DOMAIN}/group/{self.id}"
|
return f"{BASE_URL}/group/{self.id}"
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def followers_filter(cls, queryset, viewer):
|
def followers_filter(cls, queryset, viewer):
|
||||||
|
|
|
@ -2,18 +2,19 @@
|
||||||
from bookwyrm import activitypub
|
from bookwyrm import activitypub
|
||||||
from .activitypub_mixin import ActivitypubMixin
|
from .activitypub_mixin import ActivitypubMixin
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
from .fields import CICharField
|
from .fields import CharField
|
||||||
|
|
||||||
|
|
||||||
class Hashtag(ActivitypubMixin, BookWyrmModel):
|
class Hashtag(ActivitypubMixin, BookWyrmModel):
|
||||||
"a hashtag which can be used in statuses"
|
"a hashtag which can be used in statuses"
|
||||||
|
|
||||||
name = CICharField(
|
name = CharField(
|
||||||
max_length=256,
|
max_length=256,
|
||||||
blank=False,
|
blank=False,
|
||||||
null=False,
|
null=False,
|
||||||
activitypub_field="name",
|
activitypub_field="name",
|
||||||
deduplication_field=True,
|
deduplication_field=True,
|
||||||
|
db_collation="case_insensitive",
|
||||||
)
|
)
|
||||||
|
|
||||||
name_field = "name"
|
name_field = "name"
|
||||||
|
|
|
@ -135,8 +135,7 @@ class ParentJob(Job):
|
||||||
)
|
)
|
||||||
app.control.revoke(list(tasks))
|
app.control.revoke(list(tasks))
|
||||||
|
|
||||||
for task in self.pending_child_jobs:
|
self.pending_child_jobs.update(status=self.Status.STOPPED)
|
||||||
task.update(status=self.Status.STOPPED)
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def has_completed(self):
|
def has_completed(self):
|
||||||
|
@ -248,7 +247,7 @@ class SubTask(app.Task):
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def before_start(
|
def before_start(
|
||||||
self, task_id, args, kwargs
|
self, task_id, *args, **kwargs
|
||||||
): # pylint: disable=no-self-use, unused-argument
|
): # pylint: disable=no-self-use, unused-argument
|
||||||
"""Handler called before the task starts. Override.
|
"""Handler called before the task starts. Override.
|
||||||
|
|
||||||
|
@ -272,7 +271,7 @@ class SubTask(app.Task):
|
||||||
child_job.set_status(ChildJob.Status.ACTIVE)
|
child_job.set_status(ChildJob.Status.ACTIVE)
|
||||||
|
|
||||||
def on_success(
|
def on_success(
|
||||||
self, retval, task_id, args, kwargs
|
self, retval, task_id, *args, **kwargs
|
||||||
): # pylint: disable=no-self-use, unused-argument
|
): # pylint: disable=no-self-use, unused-argument
|
||||||
"""Run by the worker if the task executes successfully. Override.
|
"""Run by the worker if the task executes successfully. Override.
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
""" outlink data """
|
""" outlink data """
|
||||||
|
from typing import Optional, Iterable
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
from django.core.exceptions import PermissionDenied
|
from django.core.exceptions import PermissionDenied
|
||||||
|
@ -6,6 +7,7 @@ from django.db import models
|
||||||
from django.utils.translation import gettext_lazy as _
|
from django.utils.translation import gettext_lazy as _
|
||||||
|
|
||||||
from bookwyrm import activitypub
|
from bookwyrm import activitypub
|
||||||
|
from bookwyrm.utils.db import add_update_fields
|
||||||
from .activitypub_mixin import ActivitypubMixin
|
from .activitypub_mixin import ActivitypubMixin
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
from . import fields
|
from . import fields
|
||||||
|
@ -34,17 +36,19 @@ class Link(ActivitypubMixin, BookWyrmModel):
|
||||||
"""link name via the associated domain"""
|
"""link name via the associated domain"""
|
||||||
return self.domain.name
|
return self.domain.name
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, update_fields: Optional[Iterable[str]] = None, **kwargs):
|
||||||
"""create a link"""
|
"""create a link"""
|
||||||
# get or create the associated domain
|
# get or create the associated domain
|
||||||
if not self.domain:
|
if not self.domain:
|
||||||
domain = urlparse(self.url).netloc
|
domain = urlparse(self.url).hostname
|
||||||
self.domain, _ = LinkDomain.objects.get_or_create(domain=domain)
|
self.domain, _ = LinkDomain.objects.get_or_create(domain=domain)
|
||||||
|
update_fields = add_update_fields(update_fields, "domain")
|
||||||
|
|
||||||
# this is never broadcast, the owning model broadcasts an update
|
# this is never broadcast, the owning model broadcasts an update
|
||||||
if "broadcast" in kwargs:
|
if "broadcast" in kwargs:
|
||||||
del kwargs["broadcast"]
|
del kwargs["broadcast"]
|
||||||
return super().save(*args, **kwargs)
|
|
||||||
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
AvailabilityChoices = [
|
AvailabilityChoices = [
|
||||||
|
@ -88,8 +92,10 @@ class LinkDomain(BookWyrmModel):
|
||||||
return
|
return
|
||||||
raise PermissionDenied()
|
raise PermissionDenied()
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, update_fields: Optional[Iterable[str]] = None, **kwargs):
|
||||||
"""set a default name"""
|
"""set a default name"""
|
||||||
if not self.name:
|
if not self.name:
|
||||||
self.name = self.domain
|
self.name = self.domain
|
||||||
super().save(*args, **kwargs)
|
update_fields = add_update_fields(update_fields, "name")
|
||||||
|
|
||||||
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
""" make a list of books!! """
|
""" make a list of books!! """
|
||||||
|
from typing import Optional, Iterable
|
||||||
import uuid
|
import uuid
|
||||||
|
|
||||||
from django.core.exceptions import PermissionDenied
|
from django.core.exceptions import PermissionDenied
|
||||||
|
@ -7,7 +8,8 @@ from django.db.models import Q
|
||||||
from django.utils import timezone
|
from django.utils import timezone
|
||||||
|
|
||||||
from bookwyrm import activitypub
|
from bookwyrm import activitypub
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
|
from bookwyrm.utils.db import add_update_fields
|
||||||
|
|
||||||
from .activitypub_mixin import CollectionItemMixin, OrderedCollectionMixin
|
from .activitypub_mixin import CollectionItemMixin, OrderedCollectionMixin
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
|
@ -50,7 +52,7 @@ class List(OrderedCollectionMixin, BookWyrmModel):
|
||||||
|
|
||||||
def get_remote_id(self):
|
def get_remote_id(self):
|
||||||
"""don't want the user to be in there in this case"""
|
"""don't want the user to be in there in this case"""
|
||||||
return f"https://{DOMAIN}/list/{self.id}"
|
return f"{BASE_URL}/list/{self.id}"
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def collection_queryset(self):
|
def collection_queryset(self):
|
||||||
|
@ -124,11 +126,13 @@ class List(OrderedCollectionMixin, BookWyrmModel):
|
||||||
group=None, curation="closed"
|
group=None, curation="closed"
|
||||||
)
|
)
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, update_fields: Optional[Iterable[str]] = None, **kwargs):
|
||||||
"""on save, update embed_key and avoid clash with existing code"""
|
"""on save, update embed_key and avoid clash with existing code"""
|
||||||
if not self.embed_key:
|
if not self.embed_key:
|
||||||
self.embed_key = uuid.uuid4()
|
self.embed_key = uuid.uuid4()
|
||||||
super().save(*args, **kwargs)
|
update_fields = add_update_fields(update_fields, "embed_key")
|
||||||
|
|
||||||
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
class ListItem(CollectionItemMixin, BookWyrmModel):
|
class ListItem(CollectionItemMixin, BookWyrmModel):
|
||||||
|
|
|
@ -10,7 +10,7 @@ from .notification import Notification, NotificationType
|
||||||
|
|
||||||
|
|
||||||
class Move(ActivityMixin, BookWyrmModel):
|
class Move(ActivityMixin, BookWyrmModel):
|
||||||
"""migrating an activitypub user account"""
|
"""migrating an activitypub object"""
|
||||||
|
|
||||||
user = fields.ForeignKey(
|
user = fields.ForeignKey(
|
||||||
"User", on_delete=models.PROTECT, activitypub_field="actor"
|
"User", on_delete=models.PROTECT, activitypub_field="actor"
|
||||||
|
@ -48,24 +48,21 @@ class MoveUser(Move):
|
||||||
"""update user info and broadcast it"""
|
"""update user info and broadcast it"""
|
||||||
|
|
||||||
# only allow if the source is listed in the target's alsoKnownAs
|
# only allow if the source is listed in the target's alsoKnownAs
|
||||||
if self.user in self.target.also_known_as.all():
|
if self.user not in self.target.also_known_as.all():
|
||||||
self.user.also_known_as.add(self.target.id)
|
|
||||||
self.user.update_active_date()
|
|
||||||
self.user.moved_to = self.target.remote_id
|
|
||||||
self.user.save(update_fields=["moved_to"])
|
|
||||||
|
|
||||||
if self.user.local:
|
|
||||||
kwargs[
|
|
||||||
"broadcast"
|
|
||||||
] = True # Only broadcast if we are initiating the Move
|
|
||||||
|
|
||||||
super().save(*args, **kwargs)
|
|
||||||
|
|
||||||
for follower in self.user.followers.all():
|
|
||||||
if follower.local:
|
|
||||||
Notification.notify(
|
|
||||||
follower, self.user, notification_type=NotificationType.MOVE
|
|
||||||
)
|
|
||||||
|
|
||||||
else:
|
|
||||||
raise PermissionDenied()
|
raise PermissionDenied()
|
||||||
|
|
||||||
|
self.user.also_known_as.add(self.target.id)
|
||||||
|
self.user.update_active_date()
|
||||||
|
self.user.moved_to = self.target.remote_id
|
||||||
|
self.user.save(update_fields=["moved_to"])
|
||||||
|
|
||||||
|
if self.user.local:
|
||||||
|
kwargs["broadcast"] = True # Only broadcast if we are initiating the Move
|
||||||
|
|
||||||
|
super().save(*args, **kwargs)
|
||||||
|
|
||||||
|
for follower in self.user.followers.all():
|
||||||
|
if follower.local:
|
||||||
|
Notification.notify(
|
||||||
|
follower, self.user, notification_type=NotificationType.MOVE
|
||||||
|
)
|
||||||
|
|
|
@ -1,9 +1,13 @@
|
||||||
""" progress in a book """
|
""" progress in a book """
|
||||||
|
from typing import Optional, Iterable
|
||||||
|
|
||||||
from django.core import validators
|
from django.core import validators
|
||||||
from django.core.cache import cache
|
from django.core.cache import cache
|
||||||
from django.db import models
|
from django.db import models
|
||||||
from django.db.models import F, Q
|
from django.db.models import F, Q
|
||||||
|
|
||||||
|
from bookwyrm.utils.db import add_update_fields
|
||||||
|
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
|
|
||||||
|
|
||||||
|
@ -30,14 +34,17 @@ class ReadThrough(BookWyrmModel):
|
||||||
stopped_date = models.DateTimeField(blank=True, null=True)
|
stopped_date = models.DateTimeField(blank=True, null=True)
|
||||||
is_active = models.BooleanField(default=True)
|
is_active = models.BooleanField(default=True)
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, update_fields: Optional[Iterable[str]] = None, **kwargs):
|
||||||
"""update user active time"""
|
"""update user active time"""
|
||||||
cache.delete(f"latest_read_through-{self.user_id}-{self.book_id}")
|
|
||||||
self.user.update_active_date()
|
|
||||||
# an active readthrough must have an unset finish date
|
# an active readthrough must have an unset finish date
|
||||||
if self.finish_date or self.stopped_date:
|
if self.finish_date or self.stopped_date:
|
||||||
self.is_active = False
|
self.is_active = False
|
||||||
super().save(*args, **kwargs)
|
update_fields = add_update_fields(update_fields, "is_active")
|
||||||
|
|
||||||
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
|
cache.delete(f"latest_read_through-{self.user_id}-{self.book_id}")
|
||||||
|
self.user.update_active_date()
|
||||||
|
|
||||||
def create_update(self):
|
def create_update(self):
|
||||||
"""add update to the readthrough"""
|
"""add update to the readthrough"""
|
||||||
|
|
|
@ -38,14 +38,16 @@ class UserRelationship(BookWyrmModel):
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, **kwargs):
|
||||||
"""clear the template cache"""
|
"""clear the template cache"""
|
||||||
clear_cache(self.user_subject, self.user_object)
|
|
||||||
super().save(*args, **kwargs)
|
super().save(*args, **kwargs)
|
||||||
|
|
||||||
|
clear_cache(self.user_subject, self.user_object)
|
||||||
|
|
||||||
def delete(self, *args, **kwargs):
|
def delete(self, *args, **kwargs):
|
||||||
"""clear the template cache"""
|
"""clear the template cache"""
|
||||||
clear_cache(self.user_subject, self.user_object)
|
|
||||||
super().delete(*args, **kwargs)
|
super().delete(*args, **kwargs)
|
||||||
|
|
||||||
|
clear_cache(self.user_subject, self.user_object)
|
||||||
|
|
||||||
class Meta:
|
class Meta:
|
||||||
"""relationships should be unique"""
|
"""relationships should be unique"""
|
||||||
|
|
||||||
|
|
|
@ -3,7 +3,7 @@ from django.core.exceptions import PermissionDenied
|
||||||
from django.db import models
|
from django.db import models
|
||||||
from django.utils.translation import gettext_lazy as _
|
from django.utils.translation import gettext_lazy as _
|
||||||
|
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
|
|
||||||
|
|
||||||
|
@ -46,7 +46,7 @@ class Report(BookWyrmModel):
|
||||||
raise PermissionDenied()
|
raise PermissionDenied()
|
||||||
|
|
||||||
def get_remote_id(self):
|
def get_remote_id(self):
|
||||||
return f"https://{DOMAIN}/settings/reports/{self.id}"
|
return f"{BASE_URL}/settings/reports/{self.id}"
|
||||||
|
|
||||||
def comment(self, user, note):
|
def comment(self, user, note):
|
||||||
"""comment on a report"""
|
"""comment on a report"""
|
||||||
|
|
|
@ -1,13 +1,15 @@
|
||||||
""" puttin' books on shelves """
|
""" puttin' books on shelves """
|
||||||
import re
|
import re
|
||||||
|
from typing import Optional, Iterable
|
||||||
from django.core.cache import cache
|
from django.core.cache import cache
|
||||||
from django.core.exceptions import PermissionDenied
|
from django.core.exceptions import PermissionDenied
|
||||||
from django.db import models
|
from django.db import models
|
||||||
from django.utils import timezone
|
from django.utils import timezone
|
||||||
|
|
||||||
from bookwyrm import activitypub
|
from bookwyrm import activitypub
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
from bookwyrm.tasks import BROADCAST
|
from bookwyrm.tasks import BROADCAST
|
||||||
|
from bookwyrm.utils.db import add_update_fields
|
||||||
from .activitypub_mixin import CollectionItemMixin, OrderedCollectionMixin
|
from .activitypub_mixin import CollectionItemMixin, OrderedCollectionMixin
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
from . import fields
|
from . import fields
|
||||||
|
@ -44,8 +46,9 @@ class Shelf(OrderedCollectionMixin, BookWyrmModel):
|
||||||
"""set the identifier"""
|
"""set the identifier"""
|
||||||
super().save(*args, priority=priority, **kwargs)
|
super().save(*args, priority=priority, **kwargs)
|
||||||
if not self.identifier:
|
if not self.identifier:
|
||||||
|
# this needs the auto increment ID from the save() above
|
||||||
self.identifier = self.get_identifier()
|
self.identifier = self.get_identifier()
|
||||||
super().save(*args, **kwargs, broadcast=False)
|
super().save(*args, **kwargs, broadcast=False, update_fields={"identifier"})
|
||||||
|
|
||||||
def get_identifier(self):
|
def get_identifier(self):
|
||||||
"""custom-shelf-123 for the url"""
|
"""custom-shelf-123 for the url"""
|
||||||
|
@ -71,7 +74,7 @@ class Shelf(OrderedCollectionMixin, BookWyrmModel):
|
||||||
@property
|
@property
|
||||||
def local_path(self):
|
def local_path(self):
|
||||||
"""No slugs"""
|
"""No slugs"""
|
||||||
return self.get_remote_id().replace(f"https://{DOMAIN}", "")
|
return self.get_remote_id().replace(BASE_URL, "")
|
||||||
|
|
||||||
def raise_not_deletable(self, viewer):
|
def raise_not_deletable(self, viewer):
|
||||||
"""don't let anyone delete a default shelf"""
|
"""don't let anyone delete a default shelf"""
|
||||||
|
@ -100,10 +103,21 @@ class ShelfBook(CollectionItemMixin, BookWyrmModel):
|
||||||
activity_serializer = activitypub.ShelfItem
|
activity_serializer = activitypub.ShelfItem
|
||||||
collection_field = "shelf"
|
collection_field = "shelf"
|
||||||
|
|
||||||
def save(self, *args, priority=BROADCAST, **kwargs):
|
def save(
|
||||||
|
self,
|
||||||
|
*args,
|
||||||
|
priority=BROADCAST,
|
||||||
|
update_fields: Optional[Iterable[str]] = None,
|
||||||
|
**kwargs,
|
||||||
|
):
|
||||||
if not self.user:
|
if not self.user:
|
||||||
self.user = self.shelf.user
|
self.user = self.shelf.user
|
||||||
if self.id and self.user.local:
|
update_fields = add_update_fields(update_fields, "user")
|
||||||
|
|
||||||
|
is_update = self.id is not None
|
||||||
|
super().save(*args, priority=priority, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
|
if is_update and self.user.local:
|
||||||
# remove all caches related to all editions of this book
|
# remove all caches related to all editions of this book
|
||||||
cache.delete_many(
|
cache.delete_many(
|
||||||
[
|
[
|
||||||
|
@ -111,7 +125,6 @@ class ShelfBook(CollectionItemMixin, BookWyrmModel):
|
||||||
for book in self.book.parent_work.editions.all()
|
for book in self.book.parent_work.editions.all()
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
super().save(*args, priority=priority, **kwargs)
|
|
||||||
|
|
||||||
def delete(self, *args, **kwargs):
|
def delete(self, *args, **kwargs):
|
||||||
if self.id and self.user.local:
|
if self.id and self.user.local:
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
""" the particulars for this instance of BookWyrm """
|
""" the particulars for this instance of BookWyrm """
|
||||||
import datetime
|
import datetime
|
||||||
|
from typing import Optional, Iterable
|
||||||
from urllib.parse import urljoin
|
from urllib.parse import urljoin
|
||||||
import uuid
|
import uuid
|
||||||
|
|
||||||
|
@ -12,9 +13,10 @@ from model_utils import FieldTracker
|
||||||
|
|
||||||
from bookwyrm.connectors.abstract_connector import get_data
|
from bookwyrm.connectors.abstract_connector import get_data
|
||||||
from bookwyrm.preview_images import generate_site_preview_image_task
|
from bookwyrm.preview_images import generate_site_preview_image_task
|
||||||
from bookwyrm.settings import DOMAIN, ENABLE_PREVIEW_IMAGES, STATIC_FULL_URL
|
from bookwyrm.settings import BASE_URL, ENABLE_PREVIEW_IMAGES, STATIC_FULL_URL
|
||||||
from bookwyrm.settings import RELEASE_API
|
from bookwyrm.settings import RELEASE_API
|
||||||
from bookwyrm.tasks import app, MISC
|
from bookwyrm.tasks import app, MISC
|
||||||
|
from bookwyrm.utils.db import add_update_fields
|
||||||
from .base_model import BookWyrmModel, new_access_code
|
from .base_model import BookWyrmModel, new_access_code
|
||||||
from .user import User
|
from .user import User
|
||||||
from .fields import get_absolute_url
|
from .fields import get_absolute_url
|
||||||
|
@ -136,16 +138,19 @@ class SiteSettings(SiteModel):
|
||||||
return get_absolute_url(uploaded)
|
return get_absolute_url(uploaded)
|
||||||
return urljoin(STATIC_FULL_URL, default_path)
|
return urljoin(STATIC_FULL_URL, default_path)
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, update_fields: Optional[Iterable[str]] = None, **kwargs):
|
||||||
"""if require_confirm_email is disabled, make sure no users are pending,
|
"""if require_confirm_email is disabled, make sure no users are pending,
|
||||||
if enabled, make sure invite_question_text is not empty"""
|
if enabled, make sure invite_question_text is not empty"""
|
||||||
|
if not self.invite_question_text:
|
||||||
|
self.invite_question_text = "What is your favourite book?"
|
||||||
|
update_fields = add_update_fields(update_fields, "invite_question_text")
|
||||||
|
|
||||||
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
if not self.require_confirm_email:
|
if not self.require_confirm_email:
|
||||||
User.objects.filter(is_active=False, deactivation_reason="pending").update(
|
User.objects.filter(is_active=False, deactivation_reason="pending").update(
|
||||||
is_active=True, deactivation_reason=None
|
is_active=True, deactivation_reason=None
|
||||||
)
|
)
|
||||||
if not self.invite_question_text:
|
|
||||||
self.invite_question_text = "What is your favourite book?"
|
|
||||||
super().save(*args, **kwargs)
|
|
||||||
|
|
||||||
|
|
||||||
class Theme(SiteModel):
|
class Theme(SiteModel):
|
||||||
|
@ -188,7 +193,7 @@ class SiteInvite(models.Model):
|
||||||
@property
|
@property
|
||||||
def link(self):
|
def link(self):
|
||||||
"""formats the invite link"""
|
"""formats the invite link"""
|
||||||
return f"https://{DOMAIN}/invite/{self.code}"
|
return f"{BASE_URL}/invite/{self.code}"
|
||||||
|
|
||||||
|
|
||||||
class InviteRequest(BookWyrmModel):
|
class InviteRequest(BookWyrmModel):
|
||||||
|
@ -235,7 +240,7 @@ class PasswordReset(models.Model):
|
||||||
@property
|
@property
|
||||||
def link(self):
|
def link(self):
|
||||||
"""formats the invite link"""
|
"""formats the invite link"""
|
||||||
return f"https://{DOMAIN}/password-reset/{self.code}"
|
return f"{BASE_URL}/password-reset/{self.code}"
|
||||||
|
|
||||||
|
|
||||||
# pylint: disable=unused-argument
|
# pylint: disable=unused-argument
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
""" models for storing different kinds of Activities """
|
""" models for storing different kinds of Activities """
|
||||||
from dataclasses import MISSING
|
from dataclasses import MISSING
|
||||||
from typing import Optional
|
from typing import Optional, Iterable
|
||||||
import re
|
import re
|
||||||
|
|
||||||
from django.apps import apps
|
from django.apps import apps
|
||||||
|
@ -20,6 +20,7 @@ from model_utils.managers import InheritanceManager
|
||||||
from bookwyrm import activitypub
|
from bookwyrm import activitypub
|
||||||
from bookwyrm.preview_images import generate_edition_preview_image_task
|
from bookwyrm.preview_images import generate_edition_preview_image_task
|
||||||
from bookwyrm.settings import ENABLE_PREVIEW_IMAGES
|
from bookwyrm.settings import ENABLE_PREVIEW_IMAGES
|
||||||
|
from bookwyrm.utils.db import add_update_fields
|
||||||
from .activitypub_mixin import ActivitypubMixin, ActivityMixin
|
from .activitypub_mixin import ActivitypubMixin, ActivityMixin
|
||||||
from .activitypub_mixin import OrderedCollectionPageMixin
|
from .activitypub_mixin import OrderedCollectionPageMixin
|
||||||
from .base_model import BookWyrmModel
|
from .base_model import BookWyrmModel
|
||||||
|
@ -85,12 +86,13 @@ class Status(OrderedCollectionPageMixin, BookWyrmModel):
|
||||||
models.Index(fields=["thread_id"]),
|
models.Index(fields=["thread_id"]),
|
||||||
]
|
]
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, update_fields: Optional[Iterable[str]] = None, **kwargs):
|
||||||
"""save and notify"""
|
"""save and notify"""
|
||||||
if self.reply_parent:
|
if self.thread_id is None and self.reply_parent:
|
||||||
self.thread_id = self.reply_parent.thread_id or self.reply_parent_id
|
self.thread_id = self.reply_parent.thread_id or self.reply_parent_id
|
||||||
|
update_fields = add_update_fields(update_fields, "thread_id")
|
||||||
|
|
||||||
super().save(*args, **kwargs)
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
if not self.reply_parent:
|
if not self.reply_parent:
|
||||||
self.thread_id = self.id
|
self.thread_id = self.id
|
||||||
|
@ -392,10 +394,10 @@ class Quotation(BookStatus):
|
||||||
def _format_position(self) -> Optional[str]:
|
def _format_position(self) -> Optional[str]:
|
||||||
"""serialize page position"""
|
"""serialize page position"""
|
||||||
beg = self.position
|
beg = self.position
|
||||||
end = self.endposition or 0
|
end = self.endposition
|
||||||
if self.position_mode != "PG" or not beg:
|
if self.position_mode != "PG" or not beg:
|
||||||
return None
|
return None
|
||||||
return f"pp. {beg}-{end}" if end > beg else f"p. {beg}"
|
return f"pp. {beg}-{end}" if end else f"p. {beg}"
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def pure_content(self):
|
def pure_content(self):
|
||||||
|
@ -459,9 +461,10 @@ class Review(BookStatus):
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, **kwargs):
|
||||||
"""clear rating caches"""
|
"""clear rating caches"""
|
||||||
|
super().save(*args, **kwargs)
|
||||||
|
|
||||||
if self.book.parent_work:
|
if self.book.parent_work:
|
||||||
cache.delete(f"book-rating-{self.book.parent_work.id}")
|
cache.delete(f"book-rating-{self.book.parent_work.id}")
|
||||||
super().save(*args, **kwargs)
|
|
||||||
|
|
||||||
|
|
||||||
class ReviewRating(Review):
|
class ReviewRating(Review):
|
||||||
|
|
|
@ -1,28 +1,31 @@
|
||||||
""" database schema for user data """
|
""" database schema for user data """
|
||||||
|
import datetime
|
||||||
import re
|
import re
|
||||||
|
import zoneinfo
|
||||||
|
from typing import Optional, Iterable
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
from uuid import uuid4
|
from uuid import uuid4
|
||||||
|
|
||||||
from django.apps import apps
|
from django.apps import apps
|
||||||
from django.contrib.auth.models import AbstractUser
|
from django.contrib.auth.models import AbstractUser
|
||||||
from django.contrib.postgres.fields import ArrayField, CICharField
|
from django.contrib.postgres.fields import ArrayField as DjangoArrayField
|
||||||
from django.core.exceptions import PermissionDenied, ObjectDoesNotExist
|
from django.core.exceptions import PermissionDenied, ObjectDoesNotExist
|
||||||
from django.dispatch import receiver
|
from django.dispatch import receiver
|
||||||
from django.db import models, transaction, IntegrityError
|
from django.db import models, transaction, IntegrityError
|
||||||
from django.utils import timezone
|
from django.utils import timezone
|
||||||
from django.utils.translation import gettext_lazy as _
|
from django.utils.translation import gettext_lazy as _
|
||||||
from model_utils import FieldTracker
|
from model_utils import FieldTracker
|
||||||
import pytz
|
|
||||||
|
|
||||||
from bookwyrm import activitypub
|
from bookwyrm import activitypub
|
||||||
from bookwyrm.connectors import get_data, ConnectorException
|
from bookwyrm.connectors import get_data, ConnectorException
|
||||||
from bookwyrm.models.shelf import Shelf
|
from bookwyrm.models.shelf import Shelf
|
||||||
from bookwyrm.models.status import Status
|
from bookwyrm.models.status import Status
|
||||||
from bookwyrm.preview_images import generate_user_preview_image_task
|
from bookwyrm.preview_images import generate_user_preview_image_task
|
||||||
from bookwyrm.settings import DOMAIN, ENABLE_PREVIEW_IMAGES, USE_HTTPS, LANGUAGES
|
from bookwyrm.settings import BASE_URL, ENABLE_PREVIEW_IMAGES, LANGUAGES
|
||||||
from bookwyrm.signatures import create_key_pair
|
from bookwyrm.signatures import create_key_pair
|
||||||
from bookwyrm.tasks import app, MISC
|
from bookwyrm.tasks import app, MISC
|
||||||
from bookwyrm.utils import regex
|
from bookwyrm.utils import regex
|
||||||
|
from bookwyrm.utils.db import add_update_fields
|
||||||
from .activitypub_mixin import OrderedCollectionPageMixin, ActivitypubMixin
|
from .activitypub_mixin import OrderedCollectionPageMixin, ActivitypubMixin
|
||||||
from .base_model import BookWyrmModel, DeactivationReason, new_access_code
|
from .base_model import BookWyrmModel, DeactivationReason, new_access_code
|
||||||
from .federated_server import FederatedServer
|
from .federated_server import FederatedServer
|
||||||
|
@ -42,12 +45,6 @@ def get_feed_filter_choices():
|
||||||
return [f[0] for f in FeedFilterChoices]
|
return [f[0] for f in FeedFilterChoices]
|
||||||
|
|
||||||
|
|
||||||
def site_link():
|
|
||||||
"""helper for generating links to the site"""
|
|
||||||
protocol = "https" if USE_HTTPS else "http"
|
|
||||||
return f"{protocol}://{DOMAIN}"
|
|
||||||
|
|
||||||
|
|
||||||
# pylint: disable=too-many-public-methods
|
# pylint: disable=too-many-public-methods
|
||||||
class User(OrderedCollectionPageMixin, AbstractUser):
|
class User(OrderedCollectionPageMixin, AbstractUser):
|
||||||
"""a user who wants to read books"""
|
"""a user who wants to read books"""
|
||||||
|
@ -81,11 +78,12 @@ class User(OrderedCollectionPageMixin, AbstractUser):
|
||||||
summary = fields.HtmlField(null=True, blank=True)
|
summary = fields.HtmlField(null=True, blank=True)
|
||||||
local = models.BooleanField(default=False)
|
local = models.BooleanField(default=False)
|
||||||
bookwyrm_user = fields.BooleanField(default=True)
|
bookwyrm_user = fields.BooleanField(default=True)
|
||||||
localname = CICharField(
|
localname = models.CharField(
|
||||||
max_length=255,
|
max_length=255,
|
||||||
null=True,
|
null=True,
|
||||||
unique=True,
|
unique=True,
|
||||||
validators=[fields.validate_localname],
|
validators=[fields.validate_localname],
|
||||||
|
db_collation="case_insensitive",
|
||||||
)
|
)
|
||||||
# name is your display name, which you can change at will
|
# name is your display name, which you can change at will
|
||||||
name = fields.CharField(max_length=100, null=True, blank=True)
|
name = fields.CharField(max_length=100, null=True, blank=True)
|
||||||
|
@ -162,7 +160,7 @@ class User(OrderedCollectionPageMixin, AbstractUser):
|
||||||
show_guided_tour = models.BooleanField(default=True)
|
show_guided_tour = models.BooleanField(default=True)
|
||||||
|
|
||||||
# feed options
|
# feed options
|
||||||
feed_status_types = ArrayField(
|
feed_status_types = DjangoArrayField(
|
||||||
models.CharField(max_length=10, blank=False, choices=FeedFilterChoices),
|
models.CharField(max_length=10, blank=False, choices=FeedFilterChoices),
|
||||||
size=8,
|
size=8,
|
||||||
default=get_feed_filter_choices,
|
default=get_feed_filter_choices,
|
||||||
|
@ -171,8 +169,8 @@ class User(OrderedCollectionPageMixin, AbstractUser):
|
||||||
summary_keys = models.JSONField(null=True)
|
summary_keys = models.JSONField(null=True)
|
||||||
|
|
||||||
preferred_timezone = models.CharField(
|
preferred_timezone = models.CharField(
|
||||||
choices=[(str(tz), str(tz)) for tz in pytz.all_timezones],
|
choices=[(str(tz), str(tz)) for tz in sorted(zoneinfo.available_timezones())],
|
||||||
default=str(pytz.utc),
|
default=str(datetime.timezone.utc),
|
||||||
max_length=255,
|
max_length=255,
|
||||||
)
|
)
|
||||||
preferred_language = models.CharField(
|
preferred_language = models.CharField(
|
||||||
|
@ -214,8 +212,7 @@ class User(OrderedCollectionPageMixin, AbstractUser):
|
||||||
@property
|
@property
|
||||||
def confirmation_link(self):
|
def confirmation_link(self):
|
||||||
"""helper for generating confirmation links"""
|
"""helper for generating confirmation links"""
|
||||||
link = site_link()
|
return f"{BASE_URL}/confirm-email/{self.confirmation_code}"
|
||||||
return f"{link}/confirm-email/{self.confirmation_code}"
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def following_link(self):
|
def following_link(self):
|
||||||
|
@ -334,6 +331,7 @@ class User(OrderedCollectionPageMixin, AbstractUser):
|
||||||
"https://w3id.org/security/v1",
|
"https://w3id.org/security/v1",
|
||||||
{
|
{
|
||||||
"manuallyApprovesFollowers": "as:manuallyApprovesFollowers",
|
"manuallyApprovesFollowers": "as:manuallyApprovesFollowers",
|
||||||
|
"Hashtag": "as:Hashtag",
|
||||||
"schema": "http://schema.org#",
|
"schema": "http://schema.org#",
|
||||||
"PropertyValue": "schema:PropertyValue",
|
"PropertyValue": "schema:PropertyValue",
|
||||||
"value": "schema:value",
|
"value": "schema:value",
|
||||||
|
@ -343,13 +341,14 @@ class User(OrderedCollectionPageMixin, AbstractUser):
|
||||||
]
|
]
|
||||||
return activity_object
|
return activity_object
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, update_fields: Optional[Iterable[str]] = None, **kwargs):
|
||||||
"""populate fields for new local users"""
|
"""populate fields for new local users"""
|
||||||
created = not bool(self.id)
|
created = not bool(self.id)
|
||||||
if not self.local and not re.match(regex.FULL_USERNAME, self.username):
|
if not self.local and not re.match(regex.FULL_USERNAME, self.username):
|
||||||
# generate a username that uses the domain (webfinger format)
|
# generate a username that uses the domain (webfinger format)
|
||||||
actor_parts = urlparse(self.remote_id)
|
actor_parts = urlparse(self.remote_id)
|
||||||
self.username = f"{self.username}@{actor_parts.netloc}"
|
self.username = f"{self.username}@{actor_parts.hostname}"
|
||||||
|
update_fields = add_update_fields(update_fields, "username")
|
||||||
|
|
||||||
# this user already exists, no need to populate fields
|
# this user already exists, no need to populate fields
|
||||||
if not created:
|
if not created:
|
||||||
|
@ -358,26 +357,34 @@ class User(OrderedCollectionPageMixin, AbstractUser):
|
||||||
elif not self.deactivation_date:
|
elif not self.deactivation_date:
|
||||||
self.deactivation_date = timezone.now()
|
self.deactivation_date = timezone.now()
|
||||||
|
|
||||||
super().save(*args, **kwargs)
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
return
|
return
|
||||||
|
|
||||||
# this is a new remote user, we need to set their remote server field
|
# this is a new remote user, we need to set their remote server field
|
||||||
if not self.local:
|
if not self.local:
|
||||||
super().save(*args, **kwargs)
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
transaction.on_commit(lambda: set_remote_server(self.id))
|
transaction.on_commit(lambda: set_remote_server(self.id))
|
||||||
return
|
return
|
||||||
|
|
||||||
with transaction.atomic():
|
with transaction.atomic():
|
||||||
# populate fields for local users
|
# populate fields for local users
|
||||||
link = site_link()
|
self.remote_id = f"{BASE_URL}/user/{self.localname}"
|
||||||
self.remote_id = f"{link}/user/{self.localname}"
|
|
||||||
self.followers_url = f"{self.remote_id}/followers"
|
self.followers_url = f"{self.remote_id}/followers"
|
||||||
self.inbox = f"{self.remote_id}/inbox"
|
self.inbox = f"{self.remote_id}/inbox"
|
||||||
self.shared_inbox = f"{link}/inbox"
|
self.shared_inbox = f"{BASE_URL}/inbox"
|
||||||
self.outbox = f"{self.remote_id}/outbox"
|
self.outbox = f"{self.remote_id}/outbox"
|
||||||
|
|
||||||
|
update_fields = add_update_fields(
|
||||||
|
update_fields,
|
||||||
|
"remote_id",
|
||||||
|
"followers_url",
|
||||||
|
"inbox",
|
||||||
|
"shared_inbox",
|
||||||
|
"outbox",
|
||||||
|
)
|
||||||
|
|
||||||
# an id needs to be set before we can proceed with related models
|
# an id needs to be set before we can proceed with related models
|
||||||
super().save(*args, **kwargs)
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
# make users editors by default
|
# make users editors by default
|
||||||
try:
|
try:
|
||||||
|
@ -528,14 +535,19 @@ class KeyPair(ActivitypubMixin, BookWyrmModel):
|
||||||
# self.owner is set by the OneToOneField on User
|
# self.owner is set by the OneToOneField on User
|
||||||
return f"{self.owner.remote_id}/#main-key"
|
return f"{self.owner.remote_id}/#main-key"
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, update_fields: Optional[Iterable[str]] = None, **kwargs):
|
||||||
"""create a key pair"""
|
"""create a key pair"""
|
||||||
# no broadcasting happening here
|
# no broadcasting happening here
|
||||||
if "broadcast" in kwargs:
|
if "broadcast" in kwargs:
|
||||||
del kwargs["broadcast"]
|
del kwargs["broadcast"]
|
||||||
|
|
||||||
if not self.public_key:
|
if not self.public_key:
|
||||||
self.private_key, self.public_key = create_key_pair()
|
self.private_key, self.public_key = create_key_pair()
|
||||||
return super().save(*args, **kwargs)
|
update_fields = add_update_fields(
|
||||||
|
update_fields, "private_key", "public_key"
|
||||||
|
)
|
||||||
|
|
||||||
|
super().save(*args, update_fields=update_fields, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
@app.task(queue=MISC)
|
@app.task(queue=MISC)
|
||||||
|
@ -558,7 +570,7 @@ def set_remote_server(user_id, allow_external_connections=False):
|
||||||
user = User.objects.get(id=user_id)
|
user = User.objects.get(id=user_id)
|
||||||
actor_parts = urlparse(user.remote_id)
|
actor_parts = urlparse(user.remote_id)
|
||||||
federated_server = get_or_create_remote_server(
|
federated_server = get_or_create_remote_server(
|
||||||
actor_parts.netloc, allow_external_connections=allow_external_connections
|
actor_parts.hostname, allow_external_connections=allow_external_connections
|
||||||
)
|
)
|
||||||
# if we were unable to find the server, we need to create a new entry for it
|
# if we were unable to find the server, we need to create a new entry for it
|
||||||
if not federated_server:
|
if not federated_server:
|
||||||
|
|
|
@ -101,6 +101,7 @@ INSTALLED_APPS = [
|
||||||
"django.contrib.messages",
|
"django.contrib.messages",
|
||||||
"django.contrib.staticfiles",
|
"django.contrib.staticfiles",
|
||||||
"django.contrib.humanize",
|
"django.contrib.humanize",
|
||||||
|
"oauth2_provider",
|
||||||
"file_resubmit",
|
"file_resubmit",
|
||||||
"sass_processor",
|
"sass_processor",
|
||||||
"bookwyrm",
|
"bookwyrm",
|
||||||
|
@ -256,11 +257,8 @@ if env.bool("USE_DUMMY_CACHE", False):
|
||||||
else:
|
else:
|
||||||
CACHES = {
|
CACHES = {
|
||||||
"default": {
|
"default": {
|
||||||
"BACKEND": "django_redis.cache.RedisCache",
|
"BACKEND": "django.core.cache.backends.redis.RedisCache",
|
||||||
"LOCATION": REDIS_ACTIVITY_URL,
|
"LOCATION": REDIS_ACTIVITY_URL,
|
||||||
"OPTIONS": {
|
|
||||||
"CLIENT_CLASS": "django_redis.client.DefaultClient",
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
"file_resubmit": {
|
"file_resubmit": {
|
||||||
"BACKEND": "django.core.cache.backends.filebased.FileBasedCache",
|
"BACKEND": "django.core.cache.backends.filebased.FileBasedCache",
|
||||||
|
@ -346,34 +344,37 @@ TIME_ZONE = "UTC"
|
||||||
|
|
||||||
USE_I18N = True
|
USE_I18N = True
|
||||||
|
|
||||||
USE_L10N = True
|
|
||||||
|
|
||||||
USE_TZ = True
|
USE_TZ = True
|
||||||
|
|
||||||
|
|
||||||
USER_AGENT = f"BookWyrm (BookWyrm/{VERSION}; +https://{DOMAIN}/)"
|
|
||||||
|
|
||||||
# Imagekit generated thumbnails
|
# Imagekit generated thumbnails
|
||||||
ENABLE_THUMBNAIL_GENERATION = env.bool("ENABLE_THUMBNAIL_GENERATION", False)
|
ENABLE_THUMBNAIL_GENERATION = env.bool("ENABLE_THUMBNAIL_GENERATION", False)
|
||||||
IMAGEKIT_CACHEFILE_DIR = "thumbnails"
|
IMAGEKIT_CACHEFILE_DIR = "thumbnails"
|
||||||
IMAGEKIT_DEFAULT_CACHEFILE_STRATEGY = "bookwyrm.thumbnail_generation.Strategy"
|
IMAGEKIT_DEFAULT_CACHEFILE_STRATEGY = "bookwyrm.thumbnail_generation.Strategy"
|
||||||
|
|
||||||
# Static files (CSS, JavaScript, Images)
|
|
||||||
# https://docs.djangoproject.com/en/3.2/howto/static-files/
|
|
||||||
|
|
||||||
PROJECT_DIR = os.path.dirname(os.path.abspath(__file__))
|
PROJECT_DIR = os.path.dirname(os.path.abspath(__file__))
|
||||||
CSP_ADDITIONAL_HOSTS = env.list("CSP_ADDITIONAL_HOSTS", [])
|
CSP_ADDITIONAL_HOSTS = env.list("CSP_ADDITIONAL_HOSTS", [])
|
||||||
|
|
||||||
# Storage
|
|
||||||
|
|
||||||
PROTOCOL = "http"
|
PROTOCOL = "http"
|
||||||
if USE_HTTPS:
|
if USE_HTTPS:
|
||||||
PROTOCOL = "https"
|
PROTOCOL = "https"
|
||||||
SESSION_COOKIE_SECURE = True
|
SESSION_COOKIE_SECURE = True
|
||||||
CSRF_COOKIE_SECURE = True
|
CSRF_COOKIE_SECURE = True
|
||||||
|
|
||||||
|
PORT = env.int("PORT", 443 if USE_HTTPS else 80)
|
||||||
|
if (USE_HTTPS and PORT == 443) or (not USE_HTTPS and PORT == 80):
|
||||||
|
NETLOC = DOMAIN
|
||||||
|
else:
|
||||||
|
NETLOC = f"{DOMAIN}:{PORT}"
|
||||||
|
BASE_URL = f"{PROTOCOL}://{NETLOC}"
|
||||||
|
CSRF_TRUSTED_ORIGINS = [BASE_URL]
|
||||||
|
|
||||||
|
USER_AGENT = f"BookWyrm (BookWyrm/{VERSION}; +{BASE_URL})"
|
||||||
|
|
||||||
|
# Storage
|
||||||
|
|
||||||
USE_S3 = env.bool("USE_S3", False)
|
USE_S3 = env.bool("USE_S3", False)
|
||||||
USE_AZURE = env.bool("USE_AZURE", False)
|
USE_AZURE = env.bool("USE_AZURE", False)
|
||||||
|
S3_SIGNED_URL_EXPIRY = env.int("S3_SIGNED_URL_EXPIRY", 900)
|
||||||
|
|
||||||
if USE_S3:
|
if USE_S3:
|
||||||
# AWS settings
|
# AWS settings
|
||||||
|
@ -385,44 +386,117 @@ if USE_S3:
|
||||||
AWS_S3_ENDPOINT_URL = env("AWS_S3_ENDPOINT_URL", None)
|
AWS_S3_ENDPOINT_URL = env("AWS_S3_ENDPOINT_URL", None)
|
||||||
AWS_DEFAULT_ACL = "public-read"
|
AWS_DEFAULT_ACL = "public-read"
|
||||||
AWS_S3_OBJECT_PARAMETERS = {"CacheControl": "max-age=86400"}
|
AWS_S3_OBJECT_PARAMETERS = {"CacheControl": "max-age=86400"}
|
||||||
|
AWS_S3_URL_PROTOCOL = env("AWS_S3_URL_PROTOCOL", f"{PROTOCOL}:")
|
||||||
|
# Storages
|
||||||
|
STORAGES = {
|
||||||
|
"default": {
|
||||||
|
"BACKEND": "storages.backends.s3.S3Storage",
|
||||||
|
"OPTIONS": {
|
||||||
|
"location": "images",
|
||||||
|
"default_acl": "public-read",
|
||||||
|
"file_overwrite": False,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"staticfiles": {
|
||||||
|
"BACKEND": "storages.backends.s3.S3Storage",
|
||||||
|
"OPTIONS": {
|
||||||
|
"location": "static",
|
||||||
|
"default_acl": "public-read",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"exports": {
|
||||||
|
"BACKEND": "storages.backends.s3.S3Storage",
|
||||||
|
"OPTIONS": {
|
||||||
|
"location": "images",
|
||||||
|
"default_acl": None,
|
||||||
|
"file_overwrite": False,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
# S3 Static settings
|
# S3 Static settings
|
||||||
STATIC_LOCATION = "static"
|
STATIC_LOCATION = "static"
|
||||||
STATIC_URL = f"{PROTOCOL}://{AWS_S3_CUSTOM_DOMAIN}/{STATIC_LOCATION}/"
|
STATIC_URL = f"{AWS_S3_URL_PROTOCOL}//{AWS_S3_CUSTOM_DOMAIN}/{STATIC_LOCATION}/"
|
||||||
STATICFILES_STORAGE = "bookwyrm.storage_backends.StaticStorage"
|
STATIC_FULL_URL = STATIC_URL
|
||||||
# S3 Media settings
|
# S3 Media settings
|
||||||
MEDIA_LOCATION = "images"
|
MEDIA_LOCATION = "images"
|
||||||
MEDIA_URL = f"{PROTOCOL}://{AWS_S3_CUSTOM_DOMAIN}/{MEDIA_LOCATION}/"
|
MEDIA_URL = f"{AWS_S3_URL_PROTOCOL}//{AWS_S3_CUSTOM_DOMAIN}/{MEDIA_LOCATION}/"
|
||||||
MEDIA_FULL_URL = MEDIA_URL
|
MEDIA_FULL_URL = MEDIA_URL
|
||||||
STATIC_FULL_URL = STATIC_URL
|
# Content Security Policy
|
||||||
DEFAULT_FILE_STORAGE = "bookwyrm.storage_backends.ImagesStorage"
|
CSP_DEFAULT_SRC = [
|
||||||
CSP_DEFAULT_SRC = ["'self'", AWS_S3_CUSTOM_DOMAIN] + CSP_ADDITIONAL_HOSTS
|
"'self'",
|
||||||
CSP_SCRIPT_SRC = ["'self'", AWS_S3_CUSTOM_DOMAIN] + CSP_ADDITIONAL_HOSTS
|
f"{AWS_S3_URL_PROTOCOL}//{AWS_S3_CUSTOM_DOMAIN}"
|
||||||
|
if AWS_S3_CUSTOM_DOMAIN
|
||||||
|
else None,
|
||||||
|
] + CSP_ADDITIONAL_HOSTS
|
||||||
|
CSP_SCRIPT_SRC = [
|
||||||
|
"'self'",
|
||||||
|
f"{AWS_S3_URL_PROTOCOL}//{AWS_S3_CUSTOM_DOMAIN}"
|
||||||
|
if AWS_S3_CUSTOM_DOMAIN
|
||||||
|
else None,
|
||||||
|
] + CSP_ADDITIONAL_HOSTS
|
||||||
elif USE_AZURE:
|
elif USE_AZURE:
|
||||||
|
# Azure settings
|
||||||
AZURE_ACCOUNT_NAME = env("AZURE_ACCOUNT_NAME")
|
AZURE_ACCOUNT_NAME = env("AZURE_ACCOUNT_NAME")
|
||||||
AZURE_ACCOUNT_KEY = env("AZURE_ACCOUNT_KEY")
|
AZURE_ACCOUNT_KEY = env("AZURE_ACCOUNT_KEY")
|
||||||
AZURE_CONTAINER = env("AZURE_CONTAINER")
|
AZURE_CONTAINER = env("AZURE_CONTAINER")
|
||||||
AZURE_CUSTOM_DOMAIN = env("AZURE_CUSTOM_DOMAIN")
|
AZURE_CUSTOM_DOMAIN = env("AZURE_CUSTOM_DOMAIN")
|
||||||
|
# Storages
|
||||||
|
STORAGES = {
|
||||||
|
"default": {
|
||||||
|
"BACKEND": "storages.backends.azure_storage.AzureStorage",
|
||||||
|
"OPTIONS": {
|
||||||
|
"location": "images",
|
||||||
|
"overwrite_files": False,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"staticfiles": {
|
||||||
|
"BACKEND": "storages.backends.azure_storage.AzureStorage",
|
||||||
|
"OPTIONS": {
|
||||||
|
"location": "static",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"exports": {
|
||||||
|
"BACKEND": None, # not implemented yet
|
||||||
|
},
|
||||||
|
}
|
||||||
# Azure Static settings
|
# Azure Static settings
|
||||||
STATIC_LOCATION = "static"
|
STATIC_LOCATION = "static"
|
||||||
STATIC_URL = (
|
STATIC_URL = (
|
||||||
f"{PROTOCOL}://{AZURE_CUSTOM_DOMAIN}/{AZURE_CONTAINER}/{STATIC_LOCATION}/"
|
f"{PROTOCOL}://{AZURE_CUSTOM_DOMAIN}/{AZURE_CONTAINER}/{STATIC_LOCATION}/"
|
||||||
)
|
)
|
||||||
STATICFILES_STORAGE = "bookwyrm.storage_backends.AzureStaticStorage"
|
STATIC_FULL_URL = STATIC_URL
|
||||||
# Azure Media settings
|
# Azure Media settings
|
||||||
MEDIA_LOCATION = "images"
|
MEDIA_LOCATION = "images"
|
||||||
MEDIA_URL = (
|
MEDIA_URL = (
|
||||||
f"{PROTOCOL}://{AZURE_CUSTOM_DOMAIN}/{AZURE_CONTAINER}/{MEDIA_LOCATION}/"
|
f"{PROTOCOL}://{AZURE_CUSTOM_DOMAIN}/{AZURE_CONTAINER}/{MEDIA_LOCATION}/"
|
||||||
)
|
)
|
||||||
MEDIA_FULL_URL = MEDIA_URL
|
MEDIA_FULL_URL = MEDIA_URL
|
||||||
STATIC_FULL_URL = STATIC_URL
|
# Content Security Policy
|
||||||
DEFAULT_FILE_STORAGE = "bookwyrm.storage_backends.AzureImagesStorage"
|
|
||||||
CSP_DEFAULT_SRC = ["'self'", AZURE_CUSTOM_DOMAIN] + CSP_ADDITIONAL_HOSTS
|
CSP_DEFAULT_SRC = ["'self'", AZURE_CUSTOM_DOMAIN] + CSP_ADDITIONAL_HOSTS
|
||||||
CSP_SCRIPT_SRC = ["'self'", AZURE_CUSTOM_DOMAIN] + CSP_ADDITIONAL_HOSTS
|
CSP_SCRIPT_SRC = ["'self'", AZURE_CUSTOM_DOMAIN] + CSP_ADDITIONAL_HOSTS
|
||||||
else:
|
else:
|
||||||
|
# Storages
|
||||||
|
STORAGES = {
|
||||||
|
"default": {
|
||||||
|
"BACKEND": "django.core.files.storage.FileSystemStorage",
|
||||||
|
},
|
||||||
|
"staticfiles": {
|
||||||
|
"BACKEND": "django.contrib.staticfiles.storage.StaticFilesStorage",
|
||||||
|
},
|
||||||
|
"exports": {
|
||||||
|
"BACKEND": "django.core.files.storage.FileSystemStorage",
|
||||||
|
"OPTIONS": {
|
||||||
|
"location": "exports",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
# Static settings
|
||||||
STATIC_URL = "/static/"
|
STATIC_URL = "/static/"
|
||||||
|
STATIC_FULL_URL = BASE_URL + STATIC_URL
|
||||||
|
# Media settings
|
||||||
MEDIA_URL = "/images/"
|
MEDIA_URL = "/images/"
|
||||||
MEDIA_FULL_URL = f"{PROTOCOL}://{DOMAIN}{MEDIA_URL}"
|
MEDIA_FULL_URL = BASE_URL + MEDIA_URL
|
||||||
STATIC_FULL_URL = f"{PROTOCOL}://{DOMAIN}{STATIC_URL}"
|
# Content Security Policy
|
||||||
CSP_DEFAULT_SRC = ["'self'"] + CSP_ADDITIONAL_HOSTS
|
CSP_DEFAULT_SRC = ["'self'"] + CSP_ADDITIONAL_HOSTS
|
||||||
CSP_SCRIPT_SRC = ["'self'"] + CSP_ADDITIONAL_HOSTS
|
CSP_SCRIPT_SRC = ["'self'"] + CSP_ADDITIONAL_HOSTS
|
||||||
|
|
||||||
|
|
|
@ -1,63 +0,0 @@
|
||||||
"""Handles backends for storages"""
|
|
||||||
import os
|
|
||||||
from tempfile import SpooledTemporaryFile
|
|
||||||
from storages.backends.s3boto3 import S3Boto3Storage
|
|
||||||
from storages.backends.azure_storage import AzureStorage
|
|
||||||
|
|
||||||
|
|
||||||
class StaticStorage(S3Boto3Storage): # pylint: disable=abstract-method
|
|
||||||
"""Storage class for Static contents"""
|
|
||||||
|
|
||||||
location = "static"
|
|
||||||
default_acl = "public-read"
|
|
||||||
|
|
||||||
|
|
||||||
class ImagesStorage(S3Boto3Storage): # pylint: disable=abstract-method
|
|
||||||
"""Storage class for Image files"""
|
|
||||||
|
|
||||||
location = "images"
|
|
||||||
default_acl = "public-read"
|
|
||||||
file_overwrite = False
|
|
||||||
|
|
||||||
"""
|
|
||||||
This is our custom version of S3Boto3Storage that fixes a bug in
|
|
||||||
boto3 where the passed in file is closed upon upload.
|
|
||||||
From:
|
|
||||||
https://github.com/matthewwithanm/django-imagekit/issues/391#issuecomment-275367006
|
|
||||||
https://github.com/boto/boto3/issues/929
|
|
||||||
https://github.com/matthewwithanm/django-imagekit/issues/391
|
|
||||||
"""
|
|
||||||
|
|
||||||
def _save(self, name, content):
|
|
||||||
"""
|
|
||||||
We create a clone of the content file as when this is passed to
|
|
||||||
boto3 it wrongly closes the file upon upload where as the storage
|
|
||||||
backend expects it to still be open
|
|
||||||
"""
|
|
||||||
# Seek our content back to the start
|
|
||||||
content.seek(0, os.SEEK_SET)
|
|
||||||
|
|
||||||
# Create a temporary file that will write to disk after a specified
|
|
||||||
# size. This file will be automatically deleted when closed by
|
|
||||||
# boto3 or after exiting the `with` statement if the boto3 is fixed
|
|
||||||
with SpooledTemporaryFile() as content_autoclose:
|
|
||||||
|
|
||||||
# Write our original content into our copy that will be closed by boto3
|
|
||||||
content_autoclose.write(content.read())
|
|
||||||
|
|
||||||
# Upload the object which will auto close the
|
|
||||||
# content_autoclose instance
|
|
||||||
return super()._save(name, content_autoclose)
|
|
||||||
|
|
||||||
|
|
||||||
class AzureStaticStorage(AzureStorage): # pylint: disable=abstract-method
|
|
||||||
"""Storage class for Static contents"""
|
|
||||||
|
|
||||||
location = "static"
|
|
||||||
|
|
||||||
|
|
||||||
class AzureImagesStorage(AzureStorage): # pylint: disable=abstract-method
|
|
||||||
"""Storage class for Image files"""
|
|
||||||
|
|
||||||
location = "images"
|
|
||||||
overwrite_files = False
|
|
|
@ -2,10 +2,10 @@
|
||||||
<div style="font-family: BlinkMacSystemFont,-apple-system,'Segoe UI',Roboto,Oxygen,Ubuntu,Cantarell,'Fira Sans','Droid Sans','Helvetica Neue',Helvetica,Arial,sans-serif; border-radius: 6px; background-color: #efefef; max-width: 632px;">
|
<div style="font-family: BlinkMacSystemFont,-apple-system,'Segoe UI',Roboto,Oxygen,Ubuntu,Cantarell,'Fira Sans','Droid Sans','Helvetica Neue',Helvetica,Arial,sans-serif; border-radius: 6px; background-color: #efefef; max-width: 632px;">
|
||||||
<div style="padding: 1rem; overflow: auto;">
|
<div style="padding: 1rem; overflow: auto;">
|
||||||
<div style="float: left; margin-right: 1rem;">
|
<div style="float: left; margin-right: 1rem;">
|
||||||
<a style="color: #3273dc;" href="https://{{ domain }}" style="text-decoration: none;"><img src="{{ logo }}" alt="logo" loading="lazy" decoding="async"></a>
|
<a style="color: #3273dc;" href="{{ base_url }}" style="text-decoration: none;"><img src="{{ logo }}" alt="logo" loading="lazy" decoding="async"></a>
|
||||||
</div>
|
</div>
|
||||||
<div>
|
<div>
|
||||||
<a style="color: black; text-decoration: none" href="https://{{ domain }}" style="text-decoration: none;"><strong>{{ site_name }}</strong><br>
|
<a style="color: black; text-decoration: none" href="{{ base_url }}" style="text-decoration: none;"><strong>{{ site_name }}</strong><br>
|
||||||
{{ domain }}</a>
|
{{ domain }}</a>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
@ -18,9 +18,9 @@
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div style="padding: 1rem; font-size: 0.8rem;">
|
<div style="padding: 1rem; font-size: 0.8rem;">
|
||||||
<p style="margin: 0; color: #333;">{% blocktrans %}BookWyrm hosted on <a style="color: #3273dc;" href="https://{{ domain }}">{{ site_name }}</a>{% endblocktrans %}</p>
|
<p style="margin: 0; color: #333;">{% blocktrans %}BookWyrm hosted on <a style="color: #3273dc;" href="{{ base_url }}">{{ site_name }}</a>{% endblocktrans %}</p>
|
||||||
{% if user %}
|
{% if user %}
|
||||||
<p style="margin: 0; color: #333;"><a style="color: #3273dc;" href="https://{{ domain }}{% url 'prefs-profile' %}">{% trans "Email preference" %}</a></p>
|
<p style="margin: 0; color: #333;"><a style="color: #3273dc;" href="{{ base_url }}{% url 'prefs-profile' %}">{% trans "Email preference" %}</a></p>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|
|
@ -12,6 +12,6 @@
|
||||||
<p>
|
<p>
|
||||||
{% url 'code-of-conduct' as coc_path %}
|
{% url 'code-of-conduct' as coc_path %}
|
||||||
{% url 'about' as about_path %}
|
{% url 'about' as about_path %}
|
||||||
{% blocktrans %}Learn more <a href="https://{{ domain }}{{ about_path }}">about {{ site_name }}</a>.{% endblocktrans %}
|
{% blocktrans %}Learn more <a href="{{ base_url }}{{ about_path }}">about {{ site_name }}</a>.{% endblocktrans %}
|
||||||
</p>
|
</p>
|
||||||
{% endblock %}
|
{% endblock %}
|
||||||
|
|
|
@ -5,6 +5,6 @@
|
||||||
|
|
||||||
{{ invite_link }}
|
{{ invite_link }}
|
||||||
|
|
||||||
{% blocktrans %}Learn more about {{ site_name }}:{% endblocktrans %} https://{{ domain }}{% url 'about' %}
|
{% blocktrans %}Learn more about {{ site_name }}:{% endblocktrans %} {{ base_url }}{% url 'about' %}
|
||||||
|
|
||||||
{% endblock %}
|
{% endblock %}
|
||||||
|
|
|
@ -10,6 +10,6 @@
|
||||||
<Image width="16" height="16" type="image/x-icon">{{ image }}</Image>
|
<Image width="16" height="16" type="image/x-icon">{{ image }}</Image>
|
||||||
<Url
|
<Url
|
||||||
type="text/html"
|
type="text/html"
|
||||||
template="https://{{ DOMAIN }}{% url 'search' %}?q={searchTerms}"
|
template="{{ BASE_URL }}{% url 'search' %}?q={searchTerms}"
|
||||||
/>
|
/>
|
||||||
</OpenSearchDescription>
|
</OpenSearchDescription>
|
||||||
|
|
|
@ -97,25 +97,25 @@
|
||||||
</td>
|
</td>
|
||||||
</tr>
|
</tr>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
{% for job in jobs %}
|
{% for export in jobs %}
|
||||||
<tr>
|
<tr>
|
||||||
<td>{{ job.updated_date }}</td>
|
<td>{{ export.job.updated_date }}</td>
|
||||||
<td>
|
<td>
|
||||||
<span
|
<span
|
||||||
{% if job.status == "stopped" or job.status == "failed" %}
|
{% if export.job.status == "stopped" or export.job.status == "failed" %}
|
||||||
class="tag is-danger"
|
class="tag is-danger"
|
||||||
{% elif job.status == "pending" %}
|
{% elif export.job.status == "pending" %}
|
||||||
class="tag is-warning"
|
class="tag is-warning"
|
||||||
{% elif job.complete %}
|
{% elif export.job.complete %}
|
||||||
class="tag"
|
class="tag"
|
||||||
{% else %}
|
{% else %}
|
||||||
class="tag is-success"
|
class="tag is-success"
|
||||||
{% endif %}
|
{% endif %}
|
||||||
>
|
>
|
||||||
{% if job.status %}
|
{% if export.job.status %}
|
||||||
{{ job.status }}
|
{{ export.job.status }}
|
||||||
{{ job.status_display }}
|
{{ export.job.status_display }}
|
||||||
{% elif job.complete %}
|
{% elif export.job.complete %}
|
||||||
{% trans "Complete" %}
|
{% trans "Complete" %}
|
||||||
{% else %}
|
{% else %}
|
||||||
{% trans "Active" %}
|
{% trans "Active" %}
|
||||||
|
@ -123,18 +123,20 @@
|
||||||
</span>
|
</span>
|
||||||
</td>
|
</td>
|
||||||
<td>
|
<td>
|
||||||
<span>{{ job.export_data|get_file_size }}</span>
|
{% if export.size %}
|
||||||
|
<span>{{ export.size|get_file_size }}</span>
|
||||||
|
{% endif %}
|
||||||
</td>
|
</td>
|
||||||
<td>
|
<td>
|
||||||
{% if job.complete and not job.status == "stopped" and not job.status == "failed" %}
|
{% if export.url %}
|
||||||
<p>
|
<a href="{{ export.url }}">
|
||||||
<a download="" href="/preferences/user-export/{{ job.task_id }}">
|
<span class="icon icon-download" aria-hidden="true"></span>
|
||||||
<span class="icon icon-download" aria-hidden="true"></span>
|
<span class="is-hidden-mobile">
|
||||||
<span class="is-hidden-mobile">
|
{% trans "Download your export" %}
|
||||||
{% trans "Download your export" %}
|
</span>
|
||||||
</span>
|
</a>
|
||||||
</a>
|
{% elif export.unavailable %}
|
||||||
</p>
|
{% trans "Archive is no longer available" %}
|
||||||
{% endif %}
|
{% endif %}
|
||||||
</td>
|
</td>
|
||||||
</tr>
|
</tr>
|
||||||
|
|
|
@ -157,13 +157,13 @@
|
||||||
>
|
>
|
||||||
<div class="notification is-danger is-light">
|
<div class="notification is-danger is-light">
|
||||||
<p class="my-2">{% trans "Users are currently unable to start new user exports. This is the default setting." %}</p>
|
<p class="my-2">{% trans "Users are currently unable to start new user exports. This is the default setting." %}</p>
|
||||||
{% if use_s3 %}
|
{% if use_azure %}
|
||||||
<p>{% trans "It is not currently possible to provide user exports when using s3 storage. The BookWyrm development team are working on a fix for this." %}</p>
|
<p>{% trans "It is not currently possible to provide user exports when using Azure storage." %}</p>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
</div>
|
</div>
|
||||||
{% csrf_token %}
|
{% csrf_token %}
|
||||||
<div class="control">
|
<div class="control">
|
||||||
<button type="submit" class="button is-success" {% if use_s3 %}disabled{% endif %}>
|
<button type="submit" class="button is-success" {% if use_azure %}disabled{% endif %}>
|
||||||
{% trans "Enable user exports" %}
|
{% trans "Enable user exports" %}
|
||||||
</button>
|
</button>
|
||||||
</div>
|
</div>
|
||||||
|
|
|
@ -120,7 +120,7 @@ def id_to_username(user_id):
|
||||||
"""given an arbitrary remote id, return the username"""
|
"""given an arbitrary remote id, return the username"""
|
||||||
if user_id:
|
if user_id:
|
||||||
url = urlparse(user_id)
|
url = urlparse(user_id)
|
||||||
domain = url.netloc
|
domain = url.hostname
|
||||||
parts = url.path.split("/")
|
parts = url.path.split("/")
|
||||||
name = parts[-1]
|
name = parts[-1]
|
||||||
value = f"{name}@{domain}"
|
value = f"{name}@{domain}"
|
||||||
|
@ -130,20 +130,21 @@ def id_to_username(user_id):
|
||||||
|
|
||||||
|
|
||||||
@register.filter(name="get_file_size")
|
@register.filter(name="get_file_size")
|
||||||
def get_file_size(file):
|
def get_file_size(nbytes):
|
||||||
"""display the size of a file in human readable terms"""
|
"""display the size of a file in human readable terms"""
|
||||||
|
|
||||||
try:
|
try:
|
||||||
raw_size = os.stat(file.path).st_size
|
raw_size = float(nbytes)
|
||||||
if raw_size < 1024:
|
except (ValueError, TypeError):
|
||||||
return f"{raw_size} bytes"
|
return repr(nbytes)
|
||||||
if raw_size < 1024**2:
|
|
||||||
return f"{raw_size/1024:.2f} KB"
|
if raw_size < 1024:
|
||||||
if raw_size < 1024**3:
|
return f"{raw_size} bytes"
|
||||||
return f"{raw_size/1024**2:.2f} MB"
|
if raw_size < 1024**2:
|
||||||
return f"{raw_size/1024**3:.2f} GB"
|
return f"{raw_size/1024:.2f} KB"
|
||||||
except Exception: # pylint: disable=broad-except
|
if raw_size < 1024**3:
|
||||||
return ""
|
return f"{raw_size/1024**2:.2f} MB"
|
||||||
|
return f"{raw_size/1024**3:.2f} GB"
|
||||||
|
|
||||||
|
|
||||||
@register.filter(name="get_user_permission")
|
@register.filter(name="get_user_permission")
|
||||||
|
|
|
@ -1,8 +1,7 @@
|
||||||
""" testing activitystreams """
|
""" testing activitystreams """
|
||||||
from datetime import datetime
|
from datetime import datetime, timezone
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
from django.utils import timezone
|
|
||||||
|
|
||||||
from bookwyrm import activitystreams, models
|
from bookwyrm import activitystreams, models
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
""" testing activitystreams """
|
""" testing activitystreams """
|
||||||
from datetime import datetime, timedelta
|
import datetime
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
|
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
@ -71,8 +71,8 @@ class ActivitystreamsSignals(TestCase):
|
||||||
user=self.remote_user,
|
user=self.remote_user,
|
||||||
content="hi",
|
content="hi",
|
||||||
privacy="public",
|
privacy="public",
|
||||||
created_date=datetime(2022, 5, 16, tzinfo=timezone.utc),
|
created_date=datetime.datetime(2022, 5, 16, tzinfo=datetime.timezone.utc),
|
||||||
published_date=datetime(2022, 5, 14, tzinfo=timezone.utc),
|
published_date=datetime.datetime(2022, 5, 14, tzinfo=datetime.timezone.utc),
|
||||||
)
|
)
|
||||||
with patch("bookwyrm.activitystreams.add_status_task.apply_async") as mock:
|
with patch("bookwyrm.activitystreams.add_status_task.apply_async") as mock:
|
||||||
activitystreams.add_status_on_create_command(models.Status, status, False)
|
activitystreams.add_status_on_create_command(models.Status, status, False)
|
||||||
|
@ -87,7 +87,7 @@ class ActivitystreamsSignals(TestCase):
|
||||||
user=self.remote_user,
|
user=self.remote_user,
|
||||||
content="hi",
|
content="hi",
|
||||||
privacy="public",
|
privacy="public",
|
||||||
published_date=timezone.now() - timedelta(days=1),
|
published_date=timezone.now() - datetime.timedelta(days=1),
|
||||||
)
|
)
|
||||||
with patch("bookwyrm.activitystreams.add_status_task.apply_async") as mock:
|
with patch("bookwyrm.activitystreams.add_status_task.apply_async") as mock:
|
||||||
activitystreams.add_status_on_create_command(models.Status, status, False)
|
activitystreams.add_status_on_create_command(models.Status, status, False)
|
||||||
|
|
|
@ -6,7 +6,7 @@ import responses
|
||||||
from bookwyrm import models
|
from bookwyrm import models
|
||||||
from bookwyrm.connectors import abstract_connector, ConnectorException
|
from bookwyrm.connectors import abstract_connector, ConnectorException
|
||||||
from bookwyrm.connectors.abstract_connector import Mapping, get_data
|
from bookwyrm.connectors.abstract_connector import Mapping, get_data
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
|
|
||||||
|
|
||||||
class AbstractConnector(TestCase):
|
class AbstractConnector(TestCase):
|
||||||
|
@ -86,7 +86,7 @@ class AbstractConnector(TestCase):
|
||||||
def test_get_or_create_book_existing(self):
|
def test_get_or_create_book_existing(self):
|
||||||
"""find an existing book by remote/origin id"""
|
"""find an existing book by remote/origin id"""
|
||||||
self.assertEqual(models.Book.objects.count(), 1)
|
self.assertEqual(models.Book.objects.count(), 1)
|
||||||
self.assertEqual(self.book.remote_id, f"https://{DOMAIN}/book/{self.book.id}")
|
self.assertEqual(self.book.remote_id, f"{BASE_URL}/book/{self.book.id}")
|
||||||
self.assertEqual(self.book.origin_id, "https://example.com/book/1234")
|
self.assertEqual(self.book.origin_id, "https://example.com/book/1234")
|
||||||
|
|
||||||
# dedupe by origin id
|
# dedupe by origin id
|
||||||
|
@ -95,9 +95,7 @@ class AbstractConnector(TestCase):
|
||||||
self.assertEqual(result, self.book)
|
self.assertEqual(result, self.book)
|
||||||
|
|
||||||
# dedupe by remote id
|
# dedupe by remote id
|
||||||
result = self.connector.get_or_create_book(
|
result = self.connector.get_or_create_book(f"{BASE_URL}/book/{self.book.id}")
|
||||||
f"https://{DOMAIN}/book/{self.book.id}"
|
|
||||||
)
|
|
||||||
|
|
||||||
self.assertEqual(models.Book.objects.count(), 1)
|
self.assertEqual(models.Book.objects.count(), 1)
|
||||||
self.assertEqual(result, self.book)
|
self.assertEqual(result, self.book)
|
||||||
|
|
42
bookwyrm/tests/data/ap_user_move.json
Normal file
42
bookwyrm/tests/data/ap_user_move.json
Normal file
|
@ -0,0 +1,42 @@
|
||||||
|
{
|
||||||
|
"@context": [
|
||||||
|
"https://www.w3.org/ns/activitystreams",
|
||||||
|
"https://w3id.org/security/v1",
|
||||||
|
{
|
||||||
|
"manuallyApprovesFollowers": "as:manuallyApprovesFollowers",
|
||||||
|
"schema": "http://schema.org#",
|
||||||
|
"PropertyValue": "schema:PropertyValue",
|
||||||
|
"value": "schema:value"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"id": "https://example.com/user/mouse",
|
||||||
|
"type": "Person",
|
||||||
|
"preferredUsername": "mouse",
|
||||||
|
"name": "MOUSE?? MOUSE!!",
|
||||||
|
"inbox": "https://example.com/user/mouse/inbox",
|
||||||
|
"outbox": "https://example.com/user/mouse/outbox",
|
||||||
|
"followers": "https://example.com/user/mouse/followers",
|
||||||
|
"following": "https://example.com/user/mouse/following",
|
||||||
|
"summary": "",
|
||||||
|
"publicKey": {
|
||||||
|
"id": "https://example.com/user/mouse/#main-key",
|
||||||
|
"owner": "https://example.com/user/mouse",
|
||||||
|
"publicKeyPem": "-----BEGIN PUBLIC KEY-----\nMIGfMA0GCSqGSIb3DQEBAQUAA4GNADCBiQKBgQC6QisDrjOQvkRo/MqNmSYPwqtt\nCxg/8rCW+9jKbFUKvqjTeKVotEE85122v/DCvobCCdfQuYIFdVMk+dB1xJ0iPGPg\nyU79QHY22NdV9mFKA2qtXVVxb5cxpA4PlwOHM6PM/k8B+H09OUrop2aPUAYwy+vg\n+MXyz8bAXrIS1kq6fQIDAQAB\n-----END PUBLIC KEY-----"
|
||||||
|
},
|
||||||
|
"endpoints": {
|
||||||
|
"sharedInbox": "https://example.com/inbox"
|
||||||
|
},
|
||||||
|
"bookwyrmUser": true,
|
||||||
|
"manuallyApprovesFollowers": false,
|
||||||
|
"discoverable": false,
|
||||||
|
"alsoKnownAs": [
|
||||||
|
"https://your.domain.here:4242/user/rat"
|
||||||
|
],
|
||||||
|
"devices": "https://friend.camp/users/tripofmice/collections/devices",
|
||||||
|
"tag": [],
|
||||||
|
"icon": {
|
||||||
|
"type": "Image",
|
||||||
|
"mediaType": "image/png",
|
||||||
|
"url": "https://example.com/images/avatars/AL-2-crop-50.png"
|
||||||
|
}
|
||||||
|
}
|
Binary file not shown.
|
@ -214,7 +214,7 @@
|
||||||
"attributedTo": "https://www.example.com//user/rat",
|
"attributedTo": "https://www.example.com//user/rat",
|
||||||
"content": "<p>I like it</p>",
|
"content": "<p>I like it</p>",
|
||||||
"to": [
|
"to": [
|
||||||
"https://your.domain.here/user/rat/followers"
|
"https://your.domain.here:4242/user/rat/followers"
|
||||||
],
|
],
|
||||||
"cc": [],
|
"cc": [],
|
||||||
"replies": {
|
"replies": {
|
||||||
|
@ -395,7 +395,7 @@
|
||||||
"https://local.lists/9999"
|
"https://local.lists/9999"
|
||||||
],
|
],
|
||||||
"follows": [
|
"follows": [
|
||||||
"https://your.domain.here/user/rat"
|
"https://your.domain.here:4242/user/rat"
|
||||||
],
|
],
|
||||||
"blocks": ["https://your.domain.here/user/badger"]
|
"blocks": ["https://your.domain.here:4242/user/badger"]
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,7 +2,6 @@
|
||||||
import pathlib
|
import pathlib
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
import datetime
|
import datetime
|
||||||
import pytz
|
|
||||||
|
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
|
||||||
|
@ -13,7 +12,7 @@ from bookwyrm.models.import_job import handle_imported_book
|
||||||
|
|
||||||
def make_date(*args):
|
def make_date(*args):
|
||||||
"""helper function to easily generate a date obj"""
|
"""helper function to easily generate a date obj"""
|
||||||
return datetime.datetime(*args, tzinfo=pytz.UTC)
|
return datetime.datetime(*args, tzinfo=datetime.timezone.utc)
|
||||||
|
|
||||||
|
|
||||||
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
||||||
|
|
|
@ -3,7 +3,6 @@ from collections import namedtuple
|
||||||
import pathlib
|
import pathlib
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
import datetime
|
import datetime
|
||||||
import pytz
|
|
||||||
|
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
import responses
|
import responses
|
||||||
|
@ -16,7 +15,7 @@ from bookwyrm.models.import_job import handle_imported_book
|
||||||
|
|
||||||
def make_date(*args):
|
def make_date(*args):
|
||||||
"""helper function to easily generate a date obj"""
|
"""helper function to easily generate a date obj"""
|
||||||
return datetime.datetime(*args, tzinfo=pytz.UTC)
|
return datetime.datetime(*args, tzinfo=datetime.timezone.utc)
|
||||||
|
|
||||||
|
|
||||||
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
||||||
|
|
|
@ -2,7 +2,6 @@
|
||||||
import pathlib
|
import pathlib
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
import datetime
|
import datetime
|
||||||
import pytz
|
|
||||||
|
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
|
||||||
|
@ -13,7 +12,7 @@ from bookwyrm.models.import_job import handle_imported_book
|
||||||
|
|
||||||
def make_date(*args):
|
def make_date(*args):
|
||||||
"""helper function to easily generate a date obj"""
|
"""helper function to easily generate a date obj"""
|
||||||
return datetime.datetime(*args, tzinfo=pytz.UTC)
|
return datetime.datetime(*args, tzinfo=datetime.timezone.utc)
|
||||||
|
|
||||||
|
|
||||||
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
||||||
|
|
|
@ -2,7 +2,6 @@
|
||||||
import pathlib
|
import pathlib
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
import datetime
|
import datetime
|
||||||
import pytz
|
|
||||||
|
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
|
||||||
|
@ -13,7 +12,7 @@ from bookwyrm.models.import_job import handle_imported_book
|
||||||
|
|
||||||
def make_date(*args):
|
def make_date(*args):
|
||||||
"""helper function to easily generate a date obj"""
|
"""helper function to easily generate a date obj"""
|
||||||
return datetime.datetime(*args, tzinfo=pytz.UTC)
|
return datetime.datetime(*args, tzinfo=datetime.timezone.utc)
|
||||||
|
|
||||||
|
|
||||||
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
||||||
|
|
|
@ -2,7 +2,6 @@
|
||||||
import pathlib
|
import pathlib
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
import datetime
|
import datetime
|
||||||
import pytz
|
|
||||||
|
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
|
||||||
|
@ -13,7 +12,7 @@ from bookwyrm.models.import_job import handle_imported_book
|
||||||
|
|
||||||
def make_date(*args):
|
def make_date(*args):
|
||||||
"""helper function to easily generate a date obj"""
|
"""helper function to easily generate a date obj"""
|
||||||
return datetime.datetime(*args, tzinfo=pytz.UTC)
|
return datetime.datetime(*args, tzinfo=datetime.timezone.utc)
|
||||||
|
|
||||||
|
|
||||||
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
||||||
|
|
|
@ -5,7 +5,7 @@ from django.test import TestCase
|
||||||
|
|
||||||
from bookwyrm import models
|
from bookwyrm import models
|
||||||
from bookwyrm.models import base_model
|
from bookwyrm.models import base_model
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
|
|
||||||
|
|
||||||
# pylint: disable=attribute-defined-outside-init
|
# pylint: disable=attribute-defined-outside-init
|
||||||
|
@ -44,14 +44,14 @@ class BaseModel(TestCase):
|
||||||
"""these should be generated"""
|
"""these should be generated"""
|
||||||
self.test_model.id = 1 # pylint: disable=invalid-name
|
self.test_model.id = 1 # pylint: disable=invalid-name
|
||||||
expected = self.test_model.get_remote_id()
|
expected = self.test_model.get_remote_id()
|
||||||
self.assertEqual(expected, f"https://{DOMAIN}/bookwyrmtestmodel/1")
|
self.assertEqual(expected, f"{BASE_URL}/bookwyrmtestmodel/1")
|
||||||
|
|
||||||
def test_remote_id_with_user(self):
|
def test_remote_id_with_user(self):
|
||||||
"""format of remote id when there's a user object"""
|
"""format of remote id when there's a user object"""
|
||||||
self.test_model.user = self.local_user
|
self.test_model.user = self.local_user
|
||||||
self.test_model.id = 1
|
self.test_model.id = 1
|
||||||
expected = self.test_model.get_remote_id()
|
expected = self.test_model.get_remote_id()
|
||||||
self.assertEqual(expected, f"https://{DOMAIN}/user/mouse/bookwyrmtestmodel/1")
|
self.assertEqual(expected, f"{BASE_URL}/user/mouse/bookwyrmtestmodel/1")
|
||||||
|
|
||||||
def test_set_remote_id(self):
|
def test_set_remote_id(self):
|
||||||
"""this function sets remote ids after creation"""
|
"""this function sets remote ids after creation"""
|
||||||
|
@ -60,7 +60,7 @@ class BaseModel(TestCase):
|
||||||
instance = models.Work.objects.create(title="work title")
|
instance = models.Work.objects.create(title="work title")
|
||||||
instance.remote_id = None
|
instance.remote_id = None
|
||||||
base_model.set_remote_id(None, instance, True)
|
base_model.set_remote_id(None, instance, True)
|
||||||
self.assertEqual(instance.remote_id, f"https://{DOMAIN}/book/{instance.id}")
|
self.assertEqual(instance.remote_id, f"{BASE_URL}/book/{instance.id}")
|
||||||
|
|
||||||
# shouldn't set remote_id if it's not created
|
# shouldn't set remote_id if it's not created
|
||||||
instance.remote_id = None
|
instance.remote_id = None
|
||||||
|
|
|
@ -31,7 +31,7 @@ class Book(TestCase):
|
||||||
|
|
||||||
def test_remote_id(self):
|
def test_remote_id(self):
|
||||||
"""fanciness with remote/origin ids"""
|
"""fanciness with remote/origin ids"""
|
||||||
remote_id = f"https://{settings.DOMAIN}/book/{self.work.id}"
|
remote_id = f"{settings.BASE_URL}/book/{self.work.id}"
|
||||||
self.assertEqual(self.work.get_remote_id(), remote_id)
|
self.assertEqual(self.work.get_remote_id(), remote_id)
|
||||||
self.assertEqual(self.work.remote_id, remote_id)
|
self.assertEqual(self.work.remote_id, remote_id)
|
||||||
|
|
||||||
|
|
|
@ -1,17 +1,18 @@
|
||||||
"""test bookwyrm user export functions"""
|
"""test bookwyrm user export functions"""
|
||||||
import datetime
|
import datetime
|
||||||
import json
|
import json
|
||||||
|
import pathlib
|
||||||
|
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
|
|
||||||
from django.core.serializers.json import DjangoJSONEncoder
|
|
||||||
from django.test import TestCase
|
|
||||||
from django.utils import timezone
|
from django.utils import timezone
|
||||||
|
from django.test import TestCase
|
||||||
|
|
||||||
from bookwyrm import models
|
from bookwyrm import models
|
||||||
import bookwyrm.models.bookwyrm_export_job as export_job
|
from bookwyrm.utils.tar import BookwyrmTarFile
|
||||||
|
|
||||||
|
|
||||||
class BookwyrmExport(TestCase):
|
class BookwyrmExportJob(TestCase):
|
||||||
"""testing user export functions"""
|
"""testing user export functions"""
|
||||||
|
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
|
@ -42,6 +43,11 @@ class BookwyrmExport(TestCase):
|
||||||
preferred_timezone="America/Los Angeles",
|
preferred_timezone="America/Los Angeles",
|
||||||
default_post_privacy="followers",
|
default_post_privacy="followers",
|
||||||
)
|
)
|
||||||
|
avatar_path = pathlib.Path(__file__).parent.joinpath(
|
||||||
|
"../../static/images/default_avi.jpg"
|
||||||
|
)
|
||||||
|
with open(avatar_path, "rb") as avatar_file:
|
||||||
|
self.local_user.avatar.save("mouse-avatar.jpg", avatar_file)
|
||||||
|
|
||||||
self.rat_user = models.User.objects.create_user(
|
self.rat_user = models.User.objects.create_user(
|
||||||
"rat", "rat@rat.rat", "ratword", local=True, localname="rat"
|
"rat", "rat@rat.rat", "ratword", local=True, localname="rat"
|
||||||
|
@ -87,6 +93,13 @@ class BookwyrmExport(TestCase):
|
||||||
title="Example Edition", parent_work=self.work
|
title="Example Edition", parent_work=self.work
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# edition cover
|
||||||
|
cover_path = pathlib.Path(__file__).parent.joinpath(
|
||||||
|
"../../static/images/default_avi.jpg"
|
||||||
|
)
|
||||||
|
with open(cover_path, "rb") as cover_file:
|
||||||
|
self.edition.cover.save("tèst.jpg", cover_file)
|
||||||
|
|
||||||
self.edition.authors.add(self.author)
|
self.edition.authors.add(self.author)
|
||||||
|
|
||||||
# readthrough
|
# readthrough
|
||||||
|
@ -139,91 +152,105 @@ class BookwyrmExport(TestCase):
|
||||||
book=self.edition,
|
book=self.edition,
|
||||||
)
|
)
|
||||||
|
|
||||||
def test_json_export_user_settings(self):
|
self.job = models.BookwyrmExportJob.objects.create(user=self.local_user)
|
||||||
"""Test the json export function for basic user info"""
|
|
||||||
data = export_job.json_export(self.local_user)
|
# run the first stage of the export
|
||||||
user_data = json.loads(data)
|
with patch("bookwyrm.models.bookwyrm_export_job.create_archive_task.delay"):
|
||||||
self.assertEqual(user_data["preferredUsername"], "mouse")
|
models.bookwyrm_export_job.create_export_json_task(job_id=self.job.id)
|
||||||
self.assertEqual(user_data["name"], "Mouse")
|
self.job.refresh_from_db()
|
||||||
self.assertEqual(user_data["summary"], "<p>I'm a real bookmouse</p>")
|
|
||||||
self.assertEqual(user_data["manuallyApprovesFollowers"], False)
|
def test_add_book_to_user_export_job(self):
|
||||||
self.assertEqual(user_data["hideFollows"], False)
|
"""does AddBookToUserExportJob ...add the book to the export?"""
|
||||||
self.assertEqual(user_data["discoverable"], True)
|
self.assertIsNotNone(self.job.export_json["books"])
|
||||||
self.assertEqual(user_data["settings"]["show_goal"], False)
|
self.assertEqual(len(self.job.export_json["books"]), 1)
|
||||||
self.assertEqual(user_data["settings"]["show_suggested_users"], False)
|
book = self.job.export_json["books"][0]
|
||||||
|
|
||||||
|
self.assertEqual(book["work"]["id"], self.work.remote_id)
|
||||||
|
self.assertEqual(len(book["authors"]), 1)
|
||||||
|
self.assertEqual(len(book["shelves"]), 1)
|
||||||
|
self.assertEqual(len(book["lists"]), 1)
|
||||||
|
self.assertEqual(len(book["comments"]), 1)
|
||||||
|
self.assertEqual(len(book["reviews"]), 1)
|
||||||
|
self.assertEqual(len(book["quotations"]), 1)
|
||||||
|
self.assertEqual(len(book["readthroughs"]), 1)
|
||||||
|
|
||||||
|
self.assertEqual(book["edition"]["id"], self.edition.remote_id)
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
user_data["settings"]["preferred_timezone"], "America/Los Angeles"
|
book["edition"]["cover"]["url"], f"images/{self.edition.cover.name}"
|
||||||
)
|
|
||||||
self.assertEqual(user_data["settings"]["default_post_privacy"], "followers")
|
|
||||||
|
|
||||||
def test_json_export_extended_user_data(self):
|
|
||||||
"""Test the json export function for other non-book user info"""
|
|
||||||
data = export_job.json_export(self.local_user)
|
|
||||||
json_data = json.loads(data)
|
|
||||||
|
|
||||||
# goal
|
|
||||||
self.assertEqual(len(json_data["goals"]), 1)
|
|
||||||
self.assertEqual(json_data["goals"][0]["goal"], 128937123)
|
|
||||||
self.assertEqual(json_data["goals"][0]["year"], timezone.now().year)
|
|
||||||
self.assertEqual(json_data["goals"][0]["privacy"], "followers")
|
|
||||||
|
|
||||||
# saved lists
|
|
||||||
self.assertEqual(len(json_data["saved_lists"]), 1)
|
|
||||||
self.assertEqual(json_data["saved_lists"][0], "https://local.lists/9999")
|
|
||||||
|
|
||||||
# follows
|
|
||||||
self.assertEqual(len(json_data["follows"]), 1)
|
|
||||||
self.assertEqual(json_data["follows"][0], "https://your.domain.here/user/rat")
|
|
||||||
# blocked users
|
|
||||||
self.assertEqual(len(json_data["blocks"]), 1)
|
|
||||||
self.assertEqual(json_data["blocks"][0], "https://your.domain.here/user/badger")
|
|
||||||
|
|
||||||
def test_json_export_books(self):
|
|
||||||
"""Test the json export function for extended user info"""
|
|
||||||
|
|
||||||
data = export_job.json_export(self.local_user)
|
|
||||||
json_data = json.loads(data)
|
|
||||||
start_date = json_data["books"][0]["readthroughs"][0]["start_date"]
|
|
||||||
|
|
||||||
self.assertEqual(len(json_data["books"]), 1)
|
|
||||||
self.assertEqual(json_data["books"][0]["edition"]["title"], "Example Edition")
|
|
||||||
self.assertEqual(len(json_data["books"][0]["authors"]), 1)
|
|
||||||
self.assertEqual(json_data["books"][0]["authors"][0]["name"], "Sam Zhu")
|
|
||||||
|
|
||||||
self.assertEqual(
|
|
||||||
f'"{start_date}"', DjangoJSONEncoder().encode(self.readthrough_start)
|
|
||||||
)
|
)
|
||||||
|
|
||||||
self.assertEqual(json_data["books"][0]["shelves"][0]["name"], "Read")
|
def test_start_export_task(self):
|
||||||
|
"""test saved list task saves initial json and data"""
|
||||||
|
self.assertIsNotNone(self.job.export_data)
|
||||||
|
self.assertIsNotNone(self.job.export_json)
|
||||||
|
self.assertEqual(self.job.export_json["name"], self.local_user.name)
|
||||||
|
|
||||||
self.assertEqual(len(json_data["books"][0]["lists"]), 1)
|
def test_export_saved_lists_task(self):
|
||||||
self.assertEqual(json_data["books"][0]["lists"][0]["name"], "My excellent list")
|
"""test export_saved_lists_task adds the saved lists"""
|
||||||
|
self.assertIsNotNone(self.job.export_json["saved_lists"])
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
json_data["books"][0]["lists"][0]["list_item"]["book"],
|
self.job.export_json["saved_lists"][0], self.saved_list.remote_id
|
||||||
self.edition.remote_id,
|
|
||||||
self.edition.id,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
self.assertEqual(len(json_data["books"][0]["reviews"]), 1)
|
def test_export_follows_task(self):
|
||||||
self.assertEqual(len(json_data["books"][0]["comments"]), 1)
|
"""test export_follows_task adds the follows"""
|
||||||
self.assertEqual(len(json_data["books"][0]["quotations"]), 1)
|
self.assertIsNotNone(self.job.export_json["follows"])
|
||||||
|
self.assertEqual(self.job.export_json["follows"][0], self.rat_user.remote_id)
|
||||||
|
|
||||||
self.assertEqual(json_data["books"][0]["reviews"][0]["name"], "my review")
|
def test_export_blocks_task(self):
|
||||||
self.assertEqual(
|
"""test export_blocks_task adds the blocks"""
|
||||||
json_data["books"][0]["reviews"][0]["content"], "<p>awesome</p>"
|
self.assertIsNotNone(self.job.export_json["blocks"])
|
||||||
)
|
self.assertEqual(self.job.export_json["blocks"][0], self.badger_user.remote_id)
|
||||||
self.assertEqual(json_data["books"][0]["reviews"][0]["rating"], 5.0)
|
|
||||||
|
|
||||||
self.assertEqual(
|
def test_export_reading_goals_task(self):
|
||||||
json_data["books"][0]["comments"][0]["content"], "<p>ok so far</p>"
|
"""test export_reading_goals_task adds the goals"""
|
||||||
)
|
self.assertIsNotNone(self.job.export_json["goals"])
|
||||||
self.assertEqual(json_data["books"][0]["comments"][0]["progress"], 15)
|
self.assertEqual(self.job.export_json["goals"][0]["goal"], 128937123)
|
||||||
self.assertEqual(json_data["books"][0]["comments"][0]["progress_mode"], "PG")
|
|
||||||
|
|
||||||
|
def test_json_export(self):
|
||||||
|
"""test json_export job adds settings"""
|
||||||
|
self.assertIsNotNone(self.job.export_json["settings"])
|
||||||
|
self.assertFalse(self.job.export_json["settings"]["show_goal"])
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
json_data["books"][0]["quotations"][0]["content"], "<p>check this out</p>"
|
self.job.export_json["settings"]["preferred_timezone"],
|
||||||
|
"America/Los Angeles",
|
||||||
)
|
)
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
json_data["books"][0]["quotations"][0]["quote"],
|
self.job.export_json["settings"]["default_post_privacy"], "followers"
|
||||||
"<p>A rose by any other name</p>",
|
|
||||||
)
|
)
|
||||||
|
self.assertFalse(self.job.export_json["settings"]["show_suggested_users"])
|
||||||
|
|
||||||
|
def test_get_books_for_user(self):
|
||||||
|
"""does get_books_for_user get all the books"""
|
||||||
|
|
||||||
|
data = models.bookwyrm_export_job.get_books_for_user(self.local_user)
|
||||||
|
|
||||||
|
self.assertEqual(len(data), 1)
|
||||||
|
self.assertEqual(data[0].title, "Example Edition")
|
||||||
|
|
||||||
|
def test_archive(self):
|
||||||
|
"""actually create the TAR file"""
|
||||||
|
models.bookwyrm_export_job.create_archive_task(job_id=self.job.id)
|
||||||
|
self.job.refresh_from_db()
|
||||||
|
|
||||||
|
with (
|
||||||
|
self.job.export_data.open("rb") as tar_file,
|
||||||
|
BookwyrmTarFile.open(mode="r", fileobj=tar_file) as tar,
|
||||||
|
):
|
||||||
|
archive_json_file = tar.extractfile("archive.json")
|
||||||
|
data = json.load(archive_json_file)
|
||||||
|
|
||||||
|
# JSON from the archive should be what we want it to be
|
||||||
|
self.assertEqual(data, self.job.export_json)
|
||||||
|
|
||||||
|
# User avatar should be present in archive
|
||||||
|
with self.local_user.avatar.open() as expected_avatar:
|
||||||
|
archive_avatar = tar.extractfile(data["icon"]["url"])
|
||||||
|
self.assertEqual(expected_avatar.read(), archive_avatar.read())
|
||||||
|
|
||||||
|
# Edition cover should be present in archive
|
||||||
|
with self.edition.cover.open() as expected_cover:
|
||||||
|
archive_cover = tar.extractfile(
|
||||||
|
data["books"][0]["edition"]["cover"]["url"]
|
||||||
|
)
|
||||||
|
self.assertEqual(expected_cover.read(), archive_cover.read())
|
||||||
|
|
|
@ -4,7 +4,7 @@ from dataclasses import dataclass
|
||||||
import datetime
|
import datetime
|
||||||
import json
|
import json
|
||||||
import pathlib
|
import pathlib
|
||||||
import re
|
from urllib.parse import urlparse
|
||||||
from typing import List
|
from typing import List
|
||||||
from unittest import expectedFailure
|
from unittest import expectedFailure
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
|
@ -22,7 +22,7 @@ from bookwyrm.activitypub.base_activity import ActivityObject
|
||||||
from bookwyrm.models import fields, User, Status, Edition
|
from bookwyrm.models import fields, User, Status, Edition
|
||||||
from bookwyrm.models.base_model import BookWyrmModel
|
from bookwyrm.models.base_model import BookWyrmModel
|
||||||
from bookwyrm.models.activitypub_mixin import ActivitypubMixin
|
from bookwyrm.models.activitypub_mixin import ActivitypubMixin
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import PROTOCOL, NETLOC
|
||||||
|
|
||||||
# pylint: disable=too-many-public-methods
|
# pylint: disable=too-many-public-methods
|
||||||
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
||||||
|
@ -427,12 +427,10 @@ class ModelFields(TestCase):
|
||||||
instance = fields.ImageField()
|
instance = fields.ImageField()
|
||||||
|
|
||||||
output = instance.field_to_activity(user.avatar)
|
output = instance.field_to_activity(user.avatar)
|
||||||
self.assertIsNotNone(
|
parsed_url = urlparse(output.url)
|
||||||
re.match(
|
self.assertEqual(parsed_url.scheme, PROTOCOL)
|
||||||
rf"https:\/\/{DOMAIN}\/.*\.jpg",
|
self.assertEqual(parsed_url.netloc, NETLOC)
|
||||||
output.url,
|
self.assertRegex(parsed_url.path, r"\.jpg$")
|
||||||
)
|
|
||||||
)
|
|
||||||
self.assertEqual(output.name, "")
|
self.assertEqual(output.name, "")
|
||||||
self.assertEqual(output.type, "Image")
|
self.assertEqual(output.type, "Image")
|
||||||
|
|
||||||
|
|
|
@ -1,10 +1,10 @@
|
||||||
""" testing models """
|
""" testing models """
|
||||||
import datetime
|
import datetime
|
||||||
|
from datetime import timezone
|
||||||
import json
|
import json
|
||||||
import pathlib
|
import pathlib
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
|
|
||||||
from django.utils import timezone
|
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
import responses
|
import responses
|
||||||
|
|
||||||
|
|
|
@ -28,7 +28,7 @@ class List(TestCase):
|
||||||
def test_remote_id(self, *_):
|
def test_remote_id(self, *_):
|
||||||
"""shelves use custom remote ids"""
|
"""shelves use custom remote ids"""
|
||||||
book_list = models.List.objects.create(name="Test List", user=self.local_user)
|
book_list = models.List.objects.create(name="Test List", user=self.local_user)
|
||||||
expected_id = f"https://{settings.DOMAIN}/list/{book_list.id}"
|
expected_id = f"{settings.BASE_URL}/list/{book_list.id}"
|
||||||
self.assertEqual(book_list.get_remote_id(), expected_id)
|
self.assertEqual(book_list.get_remote_id(), expected_id)
|
||||||
|
|
||||||
def test_to_activity(self, *_):
|
def test_to_activity(self, *_):
|
||||||
|
|
60
bookwyrm/tests/models/test_move.py
Normal file
60
bookwyrm/tests/models/test_move.py
Normal file
|
@ -0,0 +1,60 @@
|
||||||
|
""" testing move models """
|
||||||
|
from unittest.mock import patch
|
||||||
|
from django.core.exceptions import PermissionDenied
|
||||||
|
from django.test import TestCase
|
||||||
|
|
||||||
|
from bookwyrm import models
|
||||||
|
|
||||||
|
|
||||||
|
class MoveUser(TestCase):
|
||||||
|
"""move your account to another identity"""
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def setUpTestData(cls):
|
||||||
|
"""we need some users for this"""
|
||||||
|
with patch("bookwyrm.models.user.set_remote_server.delay"):
|
||||||
|
cls.target_user = models.User.objects.create_user(
|
||||||
|
"rat",
|
||||||
|
"rat@rat.com",
|
||||||
|
"ratword",
|
||||||
|
local=False,
|
||||||
|
remote_id="https://example.com/users/rat",
|
||||||
|
inbox="https://example.com/users/rat/inbox",
|
||||||
|
outbox="https://example.com/users/rat/outbox",
|
||||||
|
)
|
||||||
|
|
||||||
|
with (
|
||||||
|
patch("bookwyrm.suggested_users.rerank_suggestions_task.delay"),
|
||||||
|
patch("bookwyrm.activitystreams.populate_stream_task.delay"),
|
||||||
|
patch("bookwyrm.lists_stream.populate_lists_task.delay"),
|
||||||
|
):
|
||||||
|
cls.origin_user = models.User.objects.create_user(
|
||||||
|
"mouse", "mouse@mouse.com", "mouseword", local=True, localname="mouse"
|
||||||
|
)
|
||||||
|
cls.origin_user.remote_id = "http://local.com/user/mouse"
|
||||||
|
cls.origin_user.save(broadcast=False, update_fields=["remote_id"])
|
||||||
|
|
||||||
|
def test_user_move_unauthorized(self):
|
||||||
|
"""attempt a user move without alsoKnownAs set"""
|
||||||
|
|
||||||
|
with self.assertRaises(PermissionDenied):
|
||||||
|
models.MoveUser.objects.create(
|
||||||
|
user=self.origin_user,
|
||||||
|
object=self.origin_user.remote_id,
|
||||||
|
target=self.target_user,
|
||||||
|
)
|
||||||
|
|
||||||
|
@patch("bookwyrm.suggested_users.remove_user_task.delay")
|
||||||
|
@patch("bookwyrm.models.activitypub_mixin.broadcast_task.apply_async")
|
||||||
|
def test_user_move(self, *_):
|
||||||
|
"""move user"""
|
||||||
|
|
||||||
|
self.target_user.also_known_as.add(self.origin_user.id)
|
||||||
|
self.target_user.save(broadcast=False)
|
||||||
|
|
||||||
|
models.MoveUser.objects.create(
|
||||||
|
user=self.origin_user,
|
||||||
|
object=self.origin_user.remote_id,
|
||||||
|
target=self.target_user,
|
||||||
|
)
|
||||||
|
self.assertEqual(self.origin_user.moved_to, self.target_user.remote_id)
|
|
@ -35,7 +35,7 @@ class Shelf(TestCase):
|
||||||
shelf = models.Shelf.objects.create(
|
shelf = models.Shelf.objects.create(
|
||||||
name="Test Shelf", identifier="test-shelf", user=self.local_user
|
name="Test Shelf", identifier="test-shelf", user=self.local_user
|
||||||
)
|
)
|
||||||
expected_id = f"https://{settings.DOMAIN}/user/mouse/books/test-shelf"
|
expected_id = f"{settings.BASE_URL}/user/mouse/books/test-shelf"
|
||||||
self.assertEqual(shelf.get_remote_id(), expected_id)
|
self.assertEqual(shelf.get_remote_id(), expected_id)
|
||||||
|
|
||||||
def test_to_activity(self, *_):
|
def test_to_activity(self, *_):
|
||||||
|
|
|
@ -79,7 +79,7 @@ class SiteModels(TestCase):
|
||||||
def test_site_invite_link(self):
|
def test_site_invite_link(self):
|
||||||
"""invite link generator"""
|
"""invite link generator"""
|
||||||
invite = models.SiteInvite.objects.create(user=self.local_user, code="hello")
|
invite = models.SiteInvite.objects.create(user=self.local_user, code="hello")
|
||||||
self.assertEqual(invite.link, f"https://{settings.DOMAIN}/invite/hello")
|
self.assertEqual(invite.link, f"{settings.BASE_URL}/invite/hello")
|
||||||
|
|
||||||
def test_invite_request(self):
|
def test_invite_request(self):
|
||||||
"""someone wants an invite"""
|
"""someone wants an invite"""
|
||||||
|
@ -95,7 +95,7 @@ class SiteModels(TestCase):
|
||||||
"""password reset token"""
|
"""password reset token"""
|
||||||
token = models.PasswordReset.objects.create(user=self.local_user, code="hello")
|
token = models.PasswordReset.objects.create(user=self.local_user, code="hello")
|
||||||
self.assertTrue(token.valid())
|
self.assertTrue(token.valid())
|
||||||
self.assertEqual(token.link, f"https://{settings.DOMAIN}/password-reset/hello")
|
self.assertEqual(token.link, f"{settings.BASE_URL}/password-reset/hello")
|
||||||
|
|
||||||
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
||||||
@patch("bookwyrm.suggested_users.remove_user_task.delay")
|
@patch("bookwyrm.suggested_users.remove_user_task.delay")
|
||||||
|
|
|
@ -60,7 +60,7 @@ class Status(TestCase):
|
||||||
def test_status_generated_fields(self, *_):
|
def test_status_generated_fields(self, *_):
|
||||||
"""setting remote id"""
|
"""setting remote id"""
|
||||||
status = models.Status.objects.create(content="bleh", user=self.local_user)
|
status = models.Status.objects.create(content="bleh", user=self.local_user)
|
||||||
expected_id = f"https://{settings.DOMAIN}/user/mouse/status/{status.id}"
|
expected_id = f"{settings.BASE_URL}/user/mouse/status/{status.id}"
|
||||||
self.assertEqual(status.remote_id, expected_id)
|
self.assertEqual(status.remote_id, expected_id)
|
||||||
self.assertEqual(status.privacy, "public")
|
self.assertEqual(status.privacy, "public")
|
||||||
|
|
||||||
|
@ -151,7 +151,7 @@ class Status(TestCase):
|
||||||
self.assertEqual(activity["tag"][0]["type"], "Hashtag")
|
self.assertEqual(activity["tag"][0]["type"], "Hashtag")
|
||||||
self.assertEqual(activity["tag"][0]["name"], "#content")
|
self.assertEqual(activity["tag"][0]["name"], "#content")
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
activity["tag"][0]["href"], f"https://{settings.DOMAIN}/hashtag/{tag.id}"
|
activity["tag"][0]["href"], f"{settings.BASE_URL}/hashtag/{tag.id}"
|
||||||
)
|
)
|
||||||
|
|
||||||
def test_status_with_mention_to_activity(self, *_):
|
def test_status_with_mention_to_activity(self, *_):
|
||||||
|
@ -227,11 +227,9 @@ class Status(TestCase):
|
||||||
self.assertEqual(activity["sensitive"], False)
|
self.assertEqual(activity["sensitive"], False)
|
||||||
self.assertIsInstance(activity["attachment"], list)
|
self.assertIsInstance(activity["attachment"], list)
|
||||||
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
||||||
self.assertTrue(
|
self.assertRegex(
|
||||||
re.match(
|
activity["attachment"][0]["url"],
|
||||||
r"https:\/\/your.domain.here\/images\/covers\/test(_[A-z0-9]+)?.jpg",
|
rf"^{settings.BASE_URL}/images/covers/test(_[A-z0-9]+)?.jpg$",
|
||||||
activity["attachment"][0]["url"],
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
||||||
|
|
||||||
|
@ -263,12 +261,10 @@ class Status(TestCase):
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
||||||
# self.assertTrue(
|
self.assertRegex(
|
||||||
# re.match(
|
activity["attachment"][0]["url"],
|
||||||
# r"https:\/\/your.domain.here\/images\/covers\/test_[A-z0-9]+.jpg",
|
rf"^{settings.BASE_URL}/images/covers/test_[A-z0-9]+.jpg$",
|
||||||
# activity["attachment"][0].url,
|
)
|
||||||
# )
|
|
||||||
# )
|
|
||||||
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
||||||
|
|
||||||
def test_quotation_to_activity(self, *_):
|
def test_quotation_to_activity(self, *_):
|
||||||
|
@ -306,11 +302,9 @@ class Status(TestCase):
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
||||||
self.assertTrue(
|
self.assertRegex(
|
||||||
re.match(
|
activity["attachment"][0]["url"],
|
||||||
r"https:\/\/your.domain.here\/images\/covers\/test(_[A-z0-9]+)?.jpg",
|
rf"^{settings.BASE_URL}/images/covers/test(_[A-z0-9]+)?.jpg$",
|
||||||
activity["attachment"][0]["url"],
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
||||||
|
|
||||||
|
@ -340,8 +334,11 @@ class Status(TestCase):
|
||||||
def test_quotation_page_serialization(self, *_):
|
def test_quotation_page_serialization(self, *_):
|
||||||
"""serialization of quotation page position"""
|
"""serialization of quotation page position"""
|
||||||
tests = [
|
tests = [
|
||||||
("single pos", 7, None, "p. 7"),
|
("single pos", "7", "", "p. 7"),
|
||||||
("page range", 7, 10, "pp. 7-10"),
|
("missing beg", "", "10", None),
|
||||||
|
("page range", "7", "10", "pp. 7-10"),
|
||||||
|
("page range roman", "xv", "xvi", "pp. xv-xvi"),
|
||||||
|
("page range reverse", "14", "10", "pp. 14-10"),
|
||||||
]
|
]
|
||||||
for desc, beg, end, pages in tests:
|
for desc, beg, end, pages in tests:
|
||||||
with self.subTest(desc):
|
with self.subTest(desc):
|
||||||
|
@ -355,10 +352,12 @@ class Status(TestCase):
|
||||||
position_mode="PG",
|
position_mode="PG",
|
||||||
)
|
)
|
||||||
activity = status.to_activity(pure=True)
|
activity = status.to_activity(pure=True)
|
||||||
self.assertRegex(
|
if pages:
|
||||||
activity["content"],
|
pages_re = re.escape(pages)
|
||||||
f'^<p>"my quote"</p> <p>— <a .+</a>, {pages}</p>$',
|
expect_re = f'^<p>"my quote"</p> <p>— <a .+</a>, {pages_re}</p>$'
|
||||||
)
|
else:
|
||||||
|
expect_re = '^<p>"my quote"</p> <p>— <a .+</a></p>$'
|
||||||
|
self.assertRegex(activity["content"], expect_re)
|
||||||
|
|
||||||
def test_review_to_activity(self, *_):
|
def test_review_to_activity(self, *_):
|
||||||
"""subclass of the base model version with a "pure" serializer"""
|
"""subclass of the base model version with a "pure" serializer"""
|
||||||
|
@ -395,11 +394,9 @@ class Status(TestCase):
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["content"], "test content")
|
self.assertEqual(activity["content"], "test content")
|
||||||
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
||||||
self.assertTrue(
|
self.assertRegex(
|
||||||
re.match(
|
activity["attachment"][0]["url"],
|
||||||
r"https:\/\/your.domain.here\/images\/covers\/test_[A-z0-9]+.jpg",
|
rf"^{settings.BASE_URL}/images/covers/test_[A-z0-9]+.jpg$",
|
||||||
activity["attachment"][0]["url"],
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
||||||
|
|
||||||
|
@ -420,11 +417,9 @@ class Status(TestCase):
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["content"], "test content")
|
self.assertEqual(activity["content"], "test content")
|
||||||
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
||||||
self.assertTrue(
|
self.assertRegex(
|
||||||
re.match(
|
activity["attachment"][0]["url"],
|
||||||
r"https:\/\/your.domain.here\/images\/covers\/test_[A-z0-9]+.jpg",
|
rf"^{settings.BASE_URL}/images/covers/test_[A-z0-9]+.jpg$",
|
||||||
activity["attachment"][0]["url"],
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
||||||
|
|
||||||
|
@ -443,11 +438,9 @@ class Status(TestCase):
|
||||||
f'rated <em><a href="{self.book.remote_id}">{self.book.title}</a></em>: 3 stars',
|
f'rated <em><a href="{self.book.remote_id}">{self.book.title}</a></em>: 3 stars',
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
self.assertEqual(activity["attachment"][0]["type"], "Document")
|
||||||
self.assertTrue(
|
self.assertRegex(
|
||||||
re.match(
|
activity["attachment"][0]["url"],
|
||||||
r"https:\/\/your.domain.here\/images\/covers\/test_[A-z0-9]+.jpg",
|
rf"^{settings.BASE_URL}/images/covers/test_[A-z0-9]+.jpg$",
|
||||||
activity["attachment"][0]["url"],
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
self.assertEqual(activity["attachment"][0]["name"], "Test Edition")
|
||||||
|
|
||||||
|
|
|
@ -9,15 +9,12 @@ import responses
|
||||||
|
|
||||||
from bookwyrm import models
|
from bookwyrm import models
|
||||||
from bookwyrm.management.commands import initdb
|
from bookwyrm.management.commands import initdb
|
||||||
from bookwyrm.settings import USE_HTTPS, DOMAIN
|
from bookwyrm.settings import DOMAIN, BASE_URL
|
||||||
|
|
||||||
|
|
||||||
# pylint: disable=missing-class-docstring
|
# pylint: disable=missing-class-docstring
|
||||||
# pylint: disable=missing-function-docstring
|
# pylint: disable=missing-function-docstring
|
||||||
class User(TestCase):
|
class User(TestCase):
|
||||||
|
|
||||||
protocol = "https://" if USE_HTTPS else "http://"
|
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def setUpTestData(cls):
|
def setUpTestData(cls):
|
||||||
with (
|
with (
|
||||||
|
@ -49,11 +46,11 @@ class User(TestCase):
|
||||||
|
|
||||||
def test_computed_fields(self):
|
def test_computed_fields(self):
|
||||||
"""username instead of id here"""
|
"""username instead of id here"""
|
||||||
expected_id = f"{self.protocol}{DOMAIN}/user/mouse"
|
expected_id = f"{BASE_URL}/user/mouse"
|
||||||
self.assertEqual(self.user.remote_id, expected_id)
|
self.assertEqual(self.user.remote_id, expected_id)
|
||||||
self.assertEqual(self.user.username, f"mouse@{DOMAIN}")
|
self.assertEqual(self.user.username, f"mouse@{DOMAIN}")
|
||||||
self.assertEqual(self.user.localname, "mouse")
|
self.assertEqual(self.user.localname, "mouse")
|
||||||
self.assertEqual(self.user.shared_inbox, f"{self.protocol}{DOMAIN}/inbox")
|
self.assertEqual(self.user.shared_inbox, f"{BASE_URL}/inbox")
|
||||||
self.assertEqual(self.user.inbox, f"{expected_id}/inbox")
|
self.assertEqual(self.user.inbox, f"{expected_id}/inbox")
|
||||||
self.assertEqual(self.user.outbox, f"{expected_id}/outbox")
|
self.assertEqual(self.user.outbox, f"{expected_id}/outbox")
|
||||||
self.assertEqual(self.user.followers_url, f"{expected_id}/followers")
|
self.assertEqual(self.user.followers_url, f"{expected_id}/followers")
|
||||||
|
@ -98,6 +95,7 @@ class User(TestCase):
|
||||||
"PropertyValue": "schema:PropertyValue",
|
"PropertyValue": "schema:PropertyValue",
|
||||||
"alsoKnownAs": {"@id": "as:alsoKnownAs", "@type": "@id"},
|
"alsoKnownAs": {"@id": "as:alsoKnownAs", "@type": "@id"},
|
||||||
"manuallyApprovesFollowers": "as:manuallyApprovesFollowers",
|
"manuallyApprovesFollowers": "as:manuallyApprovesFollowers",
|
||||||
|
"Hashtag": "as:Hashtag",
|
||||||
"movedTo": {"@id": "as:movedTo", "@type": "@id"},
|
"movedTo": {"@id": "as:movedTo", "@type": "@id"},
|
||||||
"schema": "http://schema.org#",
|
"schema": "http://schema.org#",
|
||||||
"value": "schema:value",
|
"value": "schema:value",
|
||||||
|
@ -130,7 +128,7 @@ class User(TestCase):
|
||||||
patch("bookwyrm.lists_stream.populate_lists_task.delay"),
|
patch("bookwyrm.lists_stream.populate_lists_task.delay"),
|
||||||
):
|
):
|
||||||
user = models.User.objects.create_user(
|
user = models.User.objects.create_user(
|
||||||
f"test2{DOMAIN}",
|
"test2",
|
||||||
"test2@bookwyrm.test",
|
"test2@bookwyrm.test",
|
||||||
localname="test2",
|
localname="test2",
|
||||||
**user_attrs,
|
**user_attrs,
|
||||||
|
@ -145,7 +143,7 @@ class User(TestCase):
|
||||||
patch("bookwyrm.lists_stream.populate_lists_task.delay"),
|
patch("bookwyrm.lists_stream.populate_lists_task.delay"),
|
||||||
):
|
):
|
||||||
user = models.User.objects.create_user(
|
user = models.User.objects.create_user(
|
||||||
f"test1{DOMAIN}",
|
"test1",
|
||||||
"test1@bookwyrm.test",
|
"test1@bookwyrm.test",
|
||||||
localname="test1",
|
localname="test1",
|
||||||
**user_attrs,
|
**user_attrs,
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
""" style fixes and lookups for templates """
|
""" style fixes and lookups for templates """
|
||||||
from datetime import datetime
|
import datetime
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
|
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
@ -95,14 +95,18 @@ class StatusDisplayTags(TestCase):
|
||||||
|
|
||||||
def test_get_published_date(self, *_):
|
def test_get_published_date(self, *_):
|
||||||
"""date formatting"""
|
"""date formatting"""
|
||||||
date = datetime(2020, 1, 1, 0, 0, tzinfo=timezone.utc)
|
date = datetime.datetime(2020, 1, 1, 0, 0, tzinfo=datetime.timezone.utc)
|
||||||
with patch("django.utils.timezone.now") as timezone_mock:
|
with patch("django.utils.timezone.now") as timezone_mock:
|
||||||
timezone_mock.return_value = datetime(2022, 1, 1, 0, 0, tzinfo=timezone.utc)
|
timezone_mock.return_value = datetime.datetime(
|
||||||
|
2022, 1, 1, 0, 0, tzinfo=datetime.timezone.utc
|
||||||
|
)
|
||||||
result = status_display.get_published_date(date)
|
result = status_display.get_published_date(date)
|
||||||
self.assertEqual(result, "Jan. 1, 2020")
|
self.assertEqual(result, "Jan. 1, 2020")
|
||||||
|
|
||||||
date = datetime(2022, 1, 1, 0, 0, tzinfo=timezone.utc)
|
date = datetime.datetime(2022, 1, 1, 0, 0, tzinfo=datetime.timezone.utc)
|
||||||
with patch("django.utils.timezone.now") as timezone_mock:
|
with patch("django.utils.timezone.now") as timezone_mock:
|
||||||
timezone_mock.return_value = datetime(2022, 1, 8, 0, 0, tzinfo=timezone.utc)
|
timezone_mock.return_value = datetime.datetime(
|
||||||
|
2022, 1, 8, 0, 0, tzinfo=datetime.timezone.utc
|
||||||
|
)
|
||||||
result = status_display.get_published_date(date)
|
result = status_display.get_published_date(date)
|
||||||
self.assertEqual(result, "Jan 1")
|
self.assertEqual(result, "Jan 1")
|
||||||
|
|
|
@ -1,8 +1,9 @@
|
||||||
""" test searching for books """
|
""" test searching for books """
|
||||||
import datetime
|
import datetime
|
||||||
|
from datetime import timezone
|
||||||
|
|
||||||
from django.db import connection
|
from django.db import connection
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
from django.utils import timezone
|
|
||||||
|
|
||||||
from bookwyrm import book_search, models
|
from bookwyrm import book_search, models
|
||||||
from bookwyrm.connectors.abstract_connector import AbstractMinimalConnector
|
from bookwyrm.connectors.abstract_connector import AbstractMinimalConnector
|
||||||
|
|
97
bookwyrm/tests/test_merge.py
Normal file
97
bookwyrm/tests/test_merge.py
Normal file
|
@ -0,0 +1,97 @@
|
||||||
|
"""test merging Authors, Works and Editions"""
|
||||||
|
|
||||||
|
from django.test import TestCase
|
||||||
|
from django.test.client import Client
|
||||||
|
|
||||||
|
from bookwyrm import models
|
||||||
|
|
||||||
|
|
||||||
|
class MergeBookDataModel(TestCase):
|
||||||
|
"""test merging of subclasses of BookDataModel"""
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def setUpTestData(cls): # pylint: disable=invalid-name
|
||||||
|
"""shared data"""
|
||||||
|
models.SiteSettings.objects.create()
|
||||||
|
|
||||||
|
cls.jrr_tolkien = models.Author.objects.create(
|
||||||
|
name="J.R.R. Tolkien",
|
||||||
|
aliases=["JRR Tolkien", "Tolkien"],
|
||||||
|
bio="This guy wrote about hobbits and stuff.",
|
||||||
|
openlibrary_key="OL26320A",
|
||||||
|
isni="0000000121441970",
|
||||||
|
)
|
||||||
|
cls.jrr_tolkien_2 = models.Author.objects.create(
|
||||||
|
name="J.R.R. Tolkien",
|
||||||
|
aliases=["JRR Tolkien", "John Ronald Reuel Tolkien"],
|
||||||
|
openlibrary_key="OL26320A",
|
||||||
|
isni="wrong",
|
||||||
|
wikidata="Q892",
|
||||||
|
)
|
||||||
|
cls.jrr_tolkien_2_id = cls.jrr_tolkien_2.id
|
||||||
|
|
||||||
|
# perform merges
|
||||||
|
cls.jrr_tolkien_absorbed_fields = cls.jrr_tolkien_2.merge_into(cls.jrr_tolkien)
|
||||||
|
|
||||||
|
def test_merged_author(self):
|
||||||
|
"""verify merged author after merge"""
|
||||||
|
self.assertEqual(self.jrr_tolkien_2.id, None, msg="duplicate should be deleted")
|
||||||
|
|
||||||
|
def test_canonical_author(self):
|
||||||
|
"""verify canonical author data after merge"""
|
||||||
|
|
||||||
|
self.assertFalse(
|
||||||
|
self.jrr_tolkien.id is None, msg="canonical should not be deleted"
|
||||||
|
)
|
||||||
|
|
||||||
|
# identical in canonical and duplicate; should be unchanged
|
||||||
|
self.assertEqual(self.jrr_tolkien.name, "J.R.R. Tolkien")
|
||||||
|
self.assertEqual(self.jrr_tolkien.openlibrary_key, "OL26320A")
|
||||||
|
|
||||||
|
# present in canonical and absent in duplicate; should be unchanged
|
||||||
|
self.assertEqual(
|
||||||
|
self.jrr_tolkien.bio, "This guy wrote about hobbits and stuff."
|
||||||
|
)
|
||||||
|
|
||||||
|
# absent in canonical and present in duplicate; should be absorbed
|
||||||
|
self.assertEqual(self.jrr_tolkien.wikidata, "Q892")
|
||||||
|
|
||||||
|
# scalar value that is different in canonical and duplicate; should be unchanged
|
||||||
|
self.assertEqual(self.jrr_tolkien.isni, "0000000121441970")
|
||||||
|
|
||||||
|
# set value with both matching and non-matching elements; should be the
|
||||||
|
# union of canonical and duplicate
|
||||||
|
self.assertEqual(
|
||||||
|
self.jrr_tolkien.aliases,
|
||||||
|
[
|
||||||
|
"JRR Tolkien",
|
||||||
|
"Tolkien",
|
||||||
|
"John Ronald Reuel Tolkien",
|
||||||
|
],
|
||||||
|
)
|
||||||
|
|
||||||
|
def test_merged_author_redirect(self):
|
||||||
|
"""a web request for a merged author should redirect to the canonical author"""
|
||||||
|
client = Client()
|
||||||
|
response = client.get(
|
||||||
|
f"/author/{self.jrr_tolkien_2_id}/s/jrr-tolkien", follow=True
|
||||||
|
)
|
||||||
|
self.assertEqual(response.redirect_chain, [(self.jrr_tolkien.local_path, 301)])
|
||||||
|
|
||||||
|
def test_merged_author_activitypub(self):
|
||||||
|
"""an activitypub request for a merged author should return the data for
|
||||||
|
the canonical author (including the canonical id)"""
|
||||||
|
client = Client(HTTP_ACCEPT="application/json")
|
||||||
|
response = client.get(f"/author/{self.jrr_tolkien_2_id}")
|
||||||
|
self.assertEqual(response.status_code, 200)
|
||||||
|
self.assertEqual(response.json(), self.jrr_tolkien.to_activity())
|
||||||
|
|
||||||
|
def test_absorbed_fields(self):
|
||||||
|
"""reported absorbed_fields should be accurate for --dry_run"""
|
||||||
|
self.assertEqual(
|
||||||
|
self.jrr_tolkien_absorbed_fields,
|
||||||
|
{
|
||||||
|
"aliases": ["John Ronald Reuel Tolkien"],
|
||||||
|
"wikidata": "Q892",
|
||||||
|
},
|
||||||
|
)
|
|
@ -1,10 +1,10 @@
|
||||||
""" test partial_date module """
|
""" test partial_date module """
|
||||||
|
|
||||||
import datetime
|
import datetime
|
||||||
|
from datetime import timezone
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from django.core.exceptions import ValidationError
|
from django.core.exceptions import ValidationError
|
||||||
from django.utils import timezone
|
|
||||||
from django.utils import translation
|
from django.utils import translation
|
||||||
|
|
||||||
from bookwyrm.utils import partial_date
|
from bookwyrm.utils import partial_date
|
||||||
|
|
|
@ -15,7 +15,7 @@ from django.utils.http import http_date
|
||||||
|
|
||||||
from bookwyrm import models
|
from bookwyrm import models
|
||||||
from bookwyrm.activitypub import Follow
|
from bookwyrm.activitypub import Follow
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import DOMAIN, NETLOC
|
||||||
from bookwyrm.signatures import create_key_pair, make_signature, make_digest
|
from bookwyrm.signatures import create_key_pair, make_signature, make_digest
|
||||||
|
|
||||||
|
|
||||||
|
@ -72,12 +72,12 @@ class Signature(TestCase):
|
||||||
urlsplit(self.rat.inbox).path,
|
urlsplit(self.rat.inbox).path,
|
||||||
data=data,
|
data=data,
|
||||||
content_type="application/json",
|
content_type="application/json",
|
||||||
**{
|
headers={
|
||||||
"HTTP_DATE": now,
|
"date": now,
|
||||||
"HTTP_SIGNATURE": signature,
|
"signature": signature,
|
||||||
"HTTP_DIGEST": digest,
|
"digest": digest,
|
||||||
"HTTP_CONTENT_TYPE": "application/activity+json; charset=utf-8",
|
"content-type": "application/activity+json; charset=utf-8",
|
||||||
"HTTP_HOST": DOMAIN,
|
"host": NETLOC,
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
|
@ -2,6 +2,7 @@
|
||||||
import re
|
import re
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
|
||||||
|
from bookwyrm.settings import BASE_URL
|
||||||
from bookwyrm.utils import regex
|
from bookwyrm.utils import regex
|
||||||
from bookwyrm.utils.validate import validate_url_domain
|
from bookwyrm.utils.validate import validate_url_domain
|
||||||
|
|
||||||
|
@ -15,17 +16,11 @@ class TestUtils(TestCase):
|
||||||
|
|
||||||
def test_valid_url_domain(self):
|
def test_valid_url_domain(self):
|
||||||
"""Check with a valid URL"""
|
"""Check with a valid URL"""
|
||||||
self.assertEqual(
|
legit = f"{BASE_URL}/legit-book-url/"
|
||||||
validate_url_domain("https://your.domain.here/legit-book-url/"),
|
self.assertEqual(validate_url_domain(legit), legit)
|
||||||
"https://your.domain.here/legit-book-url/",
|
|
||||||
)
|
|
||||||
|
|
||||||
def test_invalid_url_domain(self):
|
def test_invalid_url_domain(self):
|
||||||
"""Check with an invalid URL"""
|
"""Check with an invalid URL"""
|
||||||
self.assertIsNone(
|
self.assertIsNone(
|
||||||
validate_url_domain("https://up-to-no-good.tld/bad-actor.exe")
|
validate_url_domain("https://up-to-no-good.tld/bad-actor.exe")
|
||||||
)
|
)
|
||||||
|
|
||||||
def test_default_url_domain(self):
|
|
||||||
"""Check with a default URL"""
|
|
||||||
self.assertEqual(validate_url_domain("/"), "/")
|
|
||||||
|
|
|
@ -35,7 +35,7 @@ def validate_html(html):
|
||||||
e for e in errors.split("\n") if not any(exclude in e for exclude in excluded)
|
e for e in errors.split("\n") if not any(exclude in e for exclude in excluded)
|
||||||
)
|
)
|
||||||
if errors:
|
if errors:
|
||||||
raise Exception(errors)
|
raise ValueError(errors)
|
||||||
|
|
||||||
validator = HtmlValidator()
|
validator = HtmlValidator()
|
||||||
# will raise exceptions
|
# will raise exceptions
|
||||||
|
@ -62,6 +62,6 @@ class HtmlValidator(HTMLParser): # pylint: disable=abstract-method
|
||||||
and "noreferrer" in value
|
and "noreferrer" in value
|
||||||
):
|
):
|
||||||
return
|
return
|
||||||
raise Exception(
|
raise ValueError(
|
||||||
'Links to a new tab must have rel="nofollow noopener noreferrer"'
|
'Links to a new tab must have rel="nofollow noopener noreferrer"'
|
||||||
)
|
)
|
||||||
|
|
|
@ -123,8 +123,8 @@ class ImportViews(TestCase):
|
||||||
"""Give people a sense of the timing"""
|
"""Give people a sense of the timing"""
|
||||||
models.ImportJob.objects.create(
|
models.ImportJob.objects.create(
|
||||||
user=self.local_user,
|
user=self.local_user,
|
||||||
created_date=datetime.datetime(2000, 1, 1),
|
created_date=datetime.datetime(2000, 1, 1, tzinfo=datetime.timezone.utc),
|
||||||
updated_date=datetime.datetime(2001, 1, 1),
|
updated_date=datetime.datetime(2001, 1, 1, tzinfo=datetime.timezone.utc),
|
||||||
status="complete",
|
status="complete",
|
||||||
complete=True,
|
complete=True,
|
||||||
mappings={},
|
mappings={},
|
||||||
|
|
|
@ -134,7 +134,10 @@ class Inbox(TestCase):
|
||||||
"""check for blocked servers"""
|
"""check for blocked servers"""
|
||||||
request = self.factory.post(
|
request = self.factory.post(
|
||||||
"",
|
"",
|
||||||
HTTP_USER_AGENT="http.rb/4.4.1 (Mastodon/3.3.0; +https://mastodon.social/)",
|
headers={
|
||||||
|
# pylint: disable-next=line-too-long
|
||||||
|
"user-agent": "http.rb/4.4.1 (Mastodon/3.3.0; +https://mastodon.social/)",
|
||||||
|
},
|
||||||
)
|
)
|
||||||
self.assertIsNone(views.inbox.raise_is_blocked_user_agent(request))
|
self.assertIsNone(views.inbox.raise_is_blocked_user_agent(request))
|
||||||
|
|
||||||
|
|
|
@ -78,7 +78,7 @@ class DeleteUserViews(TestCase):
|
||||||
form.data["password"] = "password"
|
form.data["password"] = "password"
|
||||||
request = self.factory.post("", form.data)
|
request = self.factory.post("", form.data)
|
||||||
request.user = self.local_user
|
request.user = self.local_user
|
||||||
middleware = SessionMiddleware()
|
middleware = SessionMiddleware(request)
|
||||||
middleware.process_request(request)
|
middleware.process_request(request)
|
||||||
request.session.save()
|
request.session.save()
|
||||||
|
|
||||||
|
@ -105,7 +105,7 @@ class DeleteUserViews(TestCase):
|
||||||
view = views.DeactivateUser.as_view()
|
view = views.DeactivateUser.as_view()
|
||||||
request = self.factory.post("")
|
request = self.factory.post("")
|
||||||
request.user = self.local_user
|
request.user = self.local_user
|
||||||
middleware = SessionMiddleware()
|
middleware = SessionMiddleware(request)
|
||||||
middleware.process_request(request)
|
middleware.process_request(request)
|
||||||
request.session.save()
|
request.session.save()
|
||||||
|
|
||||||
|
@ -137,7 +137,7 @@ class DeleteUserViews(TestCase):
|
||||||
form.data["password"] = "password"
|
form.data["password"] = "password"
|
||||||
request = self.factory.post("", form.data)
|
request = self.factory.post("", form.data)
|
||||||
request.user = self.local_user
|
request.user = self.local_user
|
||||||
middleware = SessionMiddleware()
|
middleware = SessionMiddleware(request)
|
||||||
middleware.process_request(request)
|
middleware.process_request(request)
|
||||||
request.session.save()
|
request.session.save()
|
||||||
|
|
||||||
|
@ -159,7 +159,7 @@ class DeleteUserViews(TestCase):
|
||||||
form.data["password"] = "password"
|
form.data["password"] = "password"
|
||||||
request = self.factory.post("", form.data)
|
request = self.factory.post("", form.data)
|
||||||
request.user = self.local_user
|
request.user = self.local_user
|
||||||
middleware = SessionMiddleware()
|
middleware = SessionMiddleware(request)
|
||||||
middleware.process_request(request)
|
middleware.process_request(request)
|
||||||
request.session.save()
|
request.session.save()
|
||||||
|
|
||||||
|
|
|
@ -42,7 +42,7 @@ class ExportUserViews(TestCase):
|
||||||
|
|
||||||
request = self.factory.post("")
|
request = self.factory.post("")
|
||||||
request.user = self.local_user
|
request.user = self.local_user
|
||||||
with patch("bookwyrm.models.bookwyrm_export_job.start_export_task.delay"):
|
with patch("bookwyrm.models.bookwyrm_export_job.BookwyrmExportJob.start_job"):
|
||||||
export = views.ExportUser.as_view()(request)
|
export = views.ExportUser.as_view()(request)
|
||||||
self.assertIsInstance(export, HttpResponse)
|
self.assertIsInstance(export, HttpResponse)
|
||||||
self.assertEqual(export.status_code, 302)
|
self.assertEqual(export.status_code, 302)
|
||||||
|
|
114
bookwyrm/tests/views/preferences/test_move.py
Normal file
114
bookwyrm/tests/views/preferences/test_move.py
Normal file
|
@ -0,0 +1,114 @@
|
||||||
|
""" test move functionality """
|
||||||
|
import json
|
||||||
|
from unittest.mock import patch
|
||||||
|
import pathlib
|
||||||
|
from django.contrib.sessions.middleware import SessionMiddleware
|
||||||
|
from django.test import TestCase
|
||||||
|
from django.test.client import RequestFactory
|
||||||
|
import responses
|
||||||
|
|
||||||
|
from bookwyrm import forms, models, views
|
||||||
|
|
||||||
|
|
||||||
|
@patch("bookwyrm.activitystreams.add_status_task.delay")
|
||||||
|
@patch("bookwyrm.suggested_users.rerank_suggestions_task.delay")
|
||||||
|
@patch("bookwyrm.activitystreams.populate_stream_task.delay")
|
||||||
|
@patch("bookwyrm.suggested_users.rerank_user_task.delay")
|
||||||
|
class ViewsHelpers(TestCase):
|
||||||
|
"""viewing and creating statuses"""
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def setUpTestData(cls):
|
||||||
|
"""we need basic test data and mocks"""
|
||||||
|
|
||||||
|
with (
|
||||||
|
patch("bookwyrm.suggested_users.rerank_suggestions_task.delay"),
|
||||||
|
patch("bookwyrm.activitystreams.populate_stream_task.delay"),
|
||||||
|
patch("bookwyrm.lists_stream.populate_lists_task.delay"),
|
||||||
|
patch("bookwyrm.suggested_users.rerank_user_task.delay"),
|
||||||
|
):
|
||||||
|
cls.local_user = models.User.objects.create_user(
|
||||||
|
"rat",
|
||||||
|
"rat@rat.com",
|
||||||
|
"ratword",
|
||||||
|
local=True,
|
||||||
|
discoverable=True,
|
||||||
|
localname="rat",
|
||||||
|
)
|
||||||
|
|
||||||
|
with (
|
||||||
|
patch("bookwyrm.models.user.set_remote_server.delay"),
|
||||||
|
patch("bookwyrm.suggested_users.rerank_user_task.delay"),
|
||||||
|
):
|
||||||
|
cls.remote_user = models.User.objects.create_user(
|
||||||
|
"mouse@example.com",
|
||||||
|
"mouse@mouse.com",
|
||||||
|
"mouseword",
|
||||||
|
local=False,
|
||||||
|
remote_id="https://example.com/user/mouse",
|
||||||
|
)
|
||||||
|
|
||||||
|
def setUp(self):
|
||||||
|
"""individual test setup"""
|
||||||
|
self.factory = RequestFactory()
|
||||||
|
datafile = pathlib.Path(__file__).parent.joinpath(
|
||||||
|
"../../data/ap_user_move.json"
|
||||||
|
)
|
||||||
|
self.userdata = json.loads(datafile.read_bytes())
|
||||||
|
del self.userdata["icon"]
|
||||||
|
|
||||||
|
@responses.activate
|
||||||
|
@patch("bookwyrm.models.user.set_remote_server.delay")
|
||||||
|
@patch("bookwyrm.suggested_users.remove_user_task.delay")
|
||||||
|
@patch("bookwyrm.models.activitypub_mixin.broadcast_task.apply_async")
|
||||||
|
def test_move_user_view(self, *_):
|
||||||
|
"""move user"""
|
||||||
|
|
||||||
|
self.assertEqual(self.remote_user.remote_id, "https://example.com/user/mouse")
|
||||||
|
self.assertIsNone(self.local_user.moved_to)
|
||||||
|
self.assertIsNone(self.remote_user.moved_to)
|
||||||
|
self.assertIsNone(self.local_user.also_known_as.first())
|
||||||
|
self.assertIsNone(self.remote_user.also_known_as.first())
|
||||||
|
|
||||||
|
username = "mouse@example.com"
|
||||||
|
wellknown = {
|
||||||
|
"subject": "acct:mouse@example.com",
|
||||||
|
"links": [
|
||||||
|
{
|
||||||
|
"rel": "self",
|
||||||
|
"type": "application/activity+json",
|
||||||
|
"href": "https://example.com/user/mouse",
|
||||||
|
}
|
||||||
|
],
|
||||||
|
}
|
||||||
|
responses.add(
|
||||||
|
responses.GET,
|
||||||
|
f"https://example.com/.well-known/webfinger?resource=acct:{username}",
|
||||||
|
json=wellknown,
|
||||||
|
status=200,
|
||||||
|
)
|
||||||
|
responses.add(
|
||||||
|
responses.GET,
|
||||||
|
"https://example.com/user/mouse",
|
||||||
|
json=self.userdata,
|
||||||
|
status=200,
|
||||||
|
)
|
||||||
|
|
||||||
|
view = views.MoveUser.as_view()
|
||||||
|
form = forms.MoveUserForm()
|
||||||
|
form.data["target"] = "mouse@example.com"
|
||||||
|
form.data["password"] = "ratword"
|
||||||
|
|
||||||
|
request = self.factory.post("", form.data)
|
||||||
|
request.user = self.local_user
|
||||||
|
middleware = SessionMiddleware(request)
|
||||||
|
middleware.process_request(request)
|
||||||
|
request.session.save()
|
||||||
|
|
||||||
|
with patch("bookwyrm.models.activitypub_mixin.broadcast_task.apply_async"):
|
||||||
|
view(request)
|
||||||
|
self.local_user.refresh_from_db()
|
||||||
|
|
||||||
|
self.assertEqual(self.local_user.also_known_as.first(), self.remote_user)
|
||||||
|
self.assertEqual(self.remote_user.also_known_as.first(), self.local_user)
|
||||||
|
self.assertEqual(self.local_user.moved_to, "https://example.com/user/mouse")
|
|
@ -1,7 +1,6 @@
|
||||||
"""testing the annual summary page"""
|
"""testing the annual summary page"""
|
||||||
from datetime import datetime
|
import datetime
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
import pytz
|
|
||||||
|
|
||||||
from django.contrib.auth.models import AnonymousUser
|
from django.contrib.auth.models import AnonymousUser
|
||||||
from django.http import Http404
|
from django.http import Http404
|
||||||
|
@ -15,7 +14,7 @@ from bookwyrm.tests.validate_html import validate_html
|
||||||
|
|
||||||
def make_date(*args):
|
def make_date(*args):
|
||||||
"""helper function to easily generate a date obj"""
|
"""helper function to easily generate a date obj"""
|
||||||
return datetime(*args, tzinfo=pytz.UTC)
|
return datetime.datetime(*args, tzinfo=datetime.timezone.utc)
|
||||||
|
|
||||||
|
|
||||||
class AnnualSummary(TestCase):
|
class AnnualSummary(TestCase):
|
||||||
|
|
|
@ -8,7 +8,7 @@ from django.test.client import RequestFactory
|
||||||
import responses
|
import responses
|
||||||
|
|
||||||
from bookwyrm import models, views
|
from bookwyrm import models, views
|
||||||
from bookwyrm.settings import USER_AGENT, DOMAIN
|
from bookwyrm.settings import USER_AGENT, BASE_URL
|
||||||
|
|
||||||
|
|
||||||
@patch("bookwyrm.activitystreams.add_status_task.delay")
|
@patch("bookwyrm.activitystreams.add_status_task.delay")
|
||||||
|
@ -113,11 +113,20 @@ class ViewsHelpers(TestCase): # pylint: disable=too-many-public-methods
|
||||||
request = self.factory.get(
|
request = self.factory.get(
|
||||||
"",
|
"",
|
||||||
{"q": "Test Book"},
|
{"q": "Test Book"},
|
||||||
HTTP_USER_AGENT="http.rb/4.4.1 (Mastodon/3.3.0; +https://mastodon.social/)",
|
headers={
|
||||||
|
# pylint: disable-next=line-too-long
|
||||||
|
"user-agent": "http.rb/4.4.1 (Mastodon/3.3.0; +https://mastodon.social/)",
|
||||||
|
},
|
||||||
)
|
)
|
||||||
self.assertFalse(views.helpers.is_bookwyrm_request(request))
|
self.assertFalse(views.helpers.is_bookwyrm_request(request))
|
||||||
|
|
||||||
request = self.factory.get("", {"q": "Test Book"}, HTTP_USER_AGENT=USER_AGENT)
|
request = self.factory.get(
|
||||||
|
"",
|
||||||
|
{"q": "Test Book"},
|
||||||
|
headers={
|
||||||
|
"user-agent": USER_AGENT,
|
||||||
|
},
|
||||||
|
)
|
||||||
self.assertTrue(views.helpers.is_bookwyrm_request(request))
|
self.assertTrue(views.helpers.is_bookwyrm_request(request))
|
||||||
|
|
||||||
def test_handle_remote_webfinger_invalid(self, *_):
|
def test_handle_remote_webfinger_invalid(self, *_):
|
||||||
|
@ -271,8 +280,12 @@ class ViewsHelpers(TestCase): # pylint: disable=too-many-public-methods
|
||||||
|
|
||||||
def test_redirect_to_referer_outside_domain(self, *_):
|
def test_redirect_to_referer_outside_domain(self, *_):
|
||||||
"""safely send people on their way"""
|
"""safely send people on their way"""
|
||||||
request = self.factory.get("/path")
|
request = self.factory.get(
|
||||||
request.META = {"HTTP_REFERER": "http://outside.domain/name"}
|
"/path",
|
||||||
|
headers={
|
||||||
|
"referer": "http://outside.domain/name",
|
||||||
|
},
|
||||||
|
)
|
||||||
result = views.helpers.redirect_to_referer(
|
result = views.helpers.redirect_to_referer(
|
||||||
request, "user-feed", self.local_user.localname
|
request, "user-feed", self.local_user.localname
|
||||||
)
|
)
|
||||||
|
@ -280,21 +293,33 @@ class ViewsHelpers(TestCase): # pylint: disable=too-many-public-methods
|
||||||
|
|
||||||
def test_redirect_to_referer_outside_domain_with_fallback(self, *_):
|
def test_redirect_to_referer_outside_domain_with_fallback(self, *_):
|
||||||
"""invalid domain with regular params for the redirect function"""
|
"""invalid domain with regular params for the redirect function"""
|
||||||
request = self.factory.get("/path")
|
request = self.factory.get(
|
||||||
request.META = {"HTTP_REFERER": "https://outside.domain/name"}
|
"/path",
|
||||||
|
headers={
|
||||||
|
"referer": "http://outside.domain/name",
|
||||||
|
},
|
||||||
|
)
|
||||||
result = views.helpers.redirect_to_referer(request)
|
result = views.helpers.redirect_to_referer(request)
|
||||||
self.assertEqual(result.url, "/")
|
self.assertEqual(result.url, "/")
|
||||||
|
|
||||||
def test_redirect_to_referer_valid_domain(self, *_):
|
def test_redirect_to_referer_valid_domain(self, *_):
|
||||||
"""redirect to within the app"""
|
"""redirect to within the app"""
|
||||||
request = self.factory.get("/path")
|
request = self.factory.get(
|
||||||
request.META = {"HTTP_REFERER": f"https://{DOMAIN}/and/a/path"}
|
"/path",
|
||||||
|
headers={
|
||||||
|
"referer": f"{BASE_URL}/and/a/path",
|
||||||
|
},
|
||||||
|
)
|
||||||
result = views.helpers.redirect_to_referer(request)
|
result = views.helpers.redirect_to_referer(request)
|
||||||
self.assertEqual(result.url, f"https://{DOMAIN}/and/a/path")
|
self.assertEqual(result.url, f"{BASE_URL}/and/a/path")
|
||||||
|
|
||||||
def test_redirect_to_referer_with_get_args(self, *_):
|
def test_redirect_to_referer_with_get_args(self, *_):
|
||||||
"""if the path has get params (like sort) they are preserved"""
|
"""if the path has get params (like sort) they are preserved"""
|
||||||
request = self.factory.get("/path")
|
request = self.factory.get(
|
||||||
request.META = {"HTTP_REFERER": f"https://{DOMAIN}/and/a/path?sort=hello"}
|
"/path",
|
||||||
|
headers={
|
||||||
|
"referer": f"{BASE_URL}/and/a/path?sort=hello",
|
||||||
|
},
|
||||||
|
)
|
||||||
result = views.helpers.redirect_to_referer(request)
|
result = views.helpers.redirect_to_referer(request)
|
||||||
self.assertEqual(result.url, f"https://{DOMAIN}/and/a/path?sort=hello")
|
self.assertEqual(result.url, f"{BASE_URL}/and/a/path?sort=hello")
|
||||||
|
|
|
@ -8,7 +8,7 @@ from django.test.client import RequestFactory
|
||||||
|
|
||||||
from bookwyrm import models, views
|
from bookwyrm import models, views
|
||||||
from bookwyrm.tests.validate_html import validate_html
|
from bookwyrm.tests.validate_html import validate_html
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
|
|
||||||
|
|
||||||
class IsbnViews(TestCase):
|
class IsbnViews(TestCase):
|
||||||
|
@ -55,7 +55,7 @@ class IsbnViews(TestCase):
|
||||||
data = json.loads(response.content)
|
data = json.loads(response.content)
|
||||||
self.assertEqual(len(data), 1)
|
self.assertEqual(len(data), 1)
|
||||||
self.assertEqual(data[0]["title"], "Test Book")
|
self.assertEqual(data[0]["title"], "Test Book")
|
||||||
self.assertEqual(data[0]["key"], f"https://{DOMAIN}/book/{self.book.id}")
|
self.assertEqual(data[0]["key"], f"{BASE_URL}/book/{self.book.id}")
|
||||||
|
|
||||||
def test_isbn_html_response(self):
|
def test_isbn_html_response(self):
|
||||||
"""searches local data only and returns book data in json format"""
|
"""searches local data only and returns book data in json format"""
|
||||||
|
|
|
@ -122,7 +122,7 @@ class OutboxView(TestCase):
|
||||||
privacy="public",
|
privacy="public",
|
||||||
)
|
)
|
||||||
|
|
||||||
request = self.factory.get("", {"page": 1}, HTTP_USER_AGENT=USER_AGENT)
|
request = self.factory.get("", {"page": 1}, headers={"user-agent": USER_AGENT})
|
||||||
result = views.Outbox.as_view()(request, "mouse")
|
result = views.Outbox.as_view()(request, "mouse")
|
||||||
|
|
||||||
data = json.loads(result.content)
|
data = json.loads(result.content)
|
||||||
|
|
|
@ -1,8 +1,7 @@
|
||||||
""" tests updating reading progress """
|
""" tests updating reading progress """
|
||||||
from datetime import datetime
|
from datetime import datetime, timezone
|
||||||
from unittest.mock import patch
|
from unittest.mock import patch
|
||||||
from django.test import TestCase, Client
|
from django.test import TestCase, Client
|
||||||
from django.utils import timezone
|
|
||||||
|
|
||||||
from bookwyrm import models
|
from bookwyrm import models
|
||||||
|
|
||||||
|
|
|
@ -10,7 +10,7 @@ from django.test.client import RequestFactory
|
||||||
|
|
||||||
from bookwyrm import models, views
|
from bookwyrm import models, views
|
||||||
from bookwyrm.book_search import SearchResult
|
from bookwyrm.book_search import SearchResult
|
||||||
from bookwyrm.settings import DOMAIN
|
from bookwyrm.settings import BASE_URL
|
||||||
from bookwyrm.tests.validate_html import validate_html
|
from bookwyrm.tests.validate_html import validate_html
|
||||||
|
|
||||||
|
|
||||||
|
@ -57,7 +57,7 @@ class Views(TestCase):
|
||||||
data = json.loads(response.content)
|
data = json.loads(response.content)
|
||||||
self.assertEqual(len(data), 1)
|
self.assertEqual(len(data), 1)
|
||||||
self.assertEqual(data[0]["title"], "Test Book")
|
self.assertEqual(data[0]["title"], "Test Book")
|
||||||
self.assertEqual(data[0]["key"], f"https://{DOMAIN}/book/{self.book.id}")
|
self.assertEqual(data[0]["key"], f"{BASE_URL}/book/{self.book.id}")
|
||||||
|
|
||||||
def test_search_no_query(self):
|
def test_search_no_query(self):
|
||||||
"""just the search page"""
|
"""just the search page"""
|
||||||
|
|
Some files were not shown because too many files have changed in this diff Show more
Loading…
Add table
Reference in a new issue