2018-12-23 13:04:54 -07:00
|
|
|
# Pleroma: A lightweight social networking server
|
2023-01-02 13:38:50 -07:00
|
|
|
# Copyright © 2017-2022 Pleroma Authors <https://pleroma.social/>
|
2018-12-23 13:04:54 -07:00
|
|
|
# SPDX-License-Identifier: AGPL-3.0-only
|
|
|
|
|
2017-03-21 02:21:52 -06:00
|
|
|
defmodule Pleroma.Object do
|
|
|
|
use Ecto.Schema
|
2019-02-09 08:16:26 -07:00
|
|
|
|
2019-12-30 02:35:41 -07:00
|
|
|
import Ecto.Query
|
|
|
|
import Ecto.Changeset
|
|
|
|
|
2019-02-09 08:16:26 -07:00
|
|
|
alias Pleroma.Activity
|
2020-05-15 12:34:46 -06:00
|
|
|
alias Pleroma.Config
|
2020-12-22 12:04:33 -07:00
|
|
|
alias Pleroma.Hashtag
|
2019-03-04 19:52:23 -07:00
|
|
|
alias Pleroma.Object
|
2018-12-03 20:17:25 -07:00
|
|
|
alias Pleroma.Object.Fetcher
|
2019-02-09 08:16:26 -07:00
|
|
|
alias Pleroma.ObjectTombstone
|
2019-03-04 19:52:23 -07:00
|
|
|
alias Pleroma.Repo
|
|
|
|
alias Pleroma.User
|
2020-05-15 12:34:46 -06:00
|
|
|
alias Pleroma.Workers.AttachmentsCleanupWorker
|
2019-02-09 08:16:26 -07:00
|
|
|
|
2019-03-22 17:34:47 -06:00
|
|
|
require Logger
|
2017-03-21 02:21:52 -06:00
|
|
|
|
2020-01-12 11:48:58 -07:00
|
|
|
@type t() :: %__MODULE__{}
|
|
|
|
|
2020-01-19 09:45:20 -07:00
|
|
|
@derive {Jason.Encoder, only: [:data]}
|
|
|
|
|
2020-12-18 09:44:46 -07:00
|
|
|
@cachex Pleroma.Config.get([:cachex, :provider], Cachex)
|
|
|
|
|
2017-03-21 02:21:52 -06:00
|
|
|
schema "objects" do
|
2018-03-30 07:01:53 -06:00
|
|
|
field(:data, :map)
|
2017-03-21 02:21:52 -06:00
|
|
|
|
2020-12-22 12:04:33 -07:00
|
|
|
many_to_many(:hashtags, Hashtag, join_through: "hashtags_objects", on_replace: :delete)
|
|
|
|
|
2017-03-21 02:21:52 -06:00
|
|
|
timestamps()
|
|
|
|
end
|
2017-03-30 10:07:01 -06:00
|
|
|
|
2019-11-18 06:56:25 -07:00
|
|
|
def with_joined_activity(query, activity_type \\ "Create", join_type \\ :inner) do
|
|
|
|
object_position = Map.get(query.aliases, :object, 0)
|
|
|
|
|
|
|
|
join(query, join_type, [{object, object_position}], a in Activity,
|
|
|
|
on:
|
|
|
|
fragment(
|
2022-07-11 13:49:58 -06:00
|
|
|
"associated_object_id(?) = (? ->> 'id') AND (?->>'type' = ?) ",
|
2019-11-18 06:56:25 -07:00
|
|
|
a.data,
|
|
|
|
object.data,
|
|
|
|
a.data,
|
|
|
|
^activity_type
|
|
|
|
),
|
|
|
|
as: :object_activity
|
|
|
|
)
|
|
|
|
end
|
|
|
|
|
2017-05-16 07:31:11 -06:00
|
|
|
def create(data) do
|
2020-12-21 12:54:26 -07:00
|
|
|
%Object{}
|
|
|
|
|> Object.change(%{data: data})
|
2018-03-30 07:01:53 -06:00
|
|
|
|> Repo.insert()
|
2017-05-16 07:31:11 -06:00
|
|
|
end
|
|
|
|
|
2017-05-09 10:11:51 -06:00
|
|
|
def change(struct, params \\ %{}) do
|
2017-11-18 18:22:07 -07:00
|
|
|
struct
|
2017-05-09 10:11:51 -06:00
|
|
|
|> cast(params, [:data])
|
|
|
|
|> validate_required([:data])
|
|
|
|
|> unique_constraint(:ap_id, name: :objects_unique_apid_index)
|
2021-02-11 09:30:21 -07:00
|
|
|
# Expecting `maybe_handle_hashtags_change/1` to run last:
|
2020-12-22 12:04:33 -07:00
|
|
|
|> maybe_handle_hashtags_change(struct)
|
|
|
|
end
|
|
|
|
|
2021-02-11 09:30:21 -07:00
|
|
|
# Note: not checking activity type (assuming non-legacy objects are associated with Create act.)
|
2020-12-22 12:04:33 -07:00
|
|
|
defp maybe_handle_hashtags_change(changeset, struct) do
|
2021-02-11 09:30:21 -07:00
|
|
|
with %Ecto.Changeset{valid?: true} <- changeset,
|
|
|
|
data_hashtags_change = get_change(changeset, :data),
|
|
|
|
{_, true} <- {:changed, hashtags_changed?(struct, data_hashtags_change)},
|
2020-12-22 12:04:33 -07:00
|
|
|
{:ok, hashtag_records} <-
|
|
|
|
data_hashtags_change
|
|
|
|
|> object_data_hashtags()
|
|
|
|
|> Hashtag.get_or_create_by_names() do
|
|
|
|
put_assoc(changeset, :hashtags, hashtag_records)
|
2020-12-21 12:54:26 -07:00
|
|
|
else
|
2021-02-11 09:30:21 -07:00
|
|
|
%{valid?: false} ->
|
2020-12-22 12:04:33 -07:00
|
|
|
changeset
|
|
|
|
|
2021-02-11 09:30:21 -07:00
|
|
|
{:changed, false} ->
|
|
|
|
changeset
|
2020-12-22 12:04:33 -07:00
|
|
|
|
2021-02-11 09:30:21 -07:00
|
|
|
{:error, _} ->
|
2020-12-22 12:04:33 -07:00
|
|
|
validate_change(changeset, :data, fn _, _ ->
|
2021-02-11 09:30:21 -07:00
|
|
|
[data: "error referencing hashtags"]
|
2020-12-22 12:04:33 -07:00
|
|
|
end)
|
2020-12-21 12:54:26 -07:00
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
defp hashtags_changed?(%Object{} = struct, %{"tag" => _} = data) do
|
|
|
|
Enum.sort(embedded_hashtags(struct)) !=
|
|
|
|
Enum.sort(object_data_hashtags(data))
|
2017-05-09 10:11:51 -06:00
|
|
|
end
|
|
|
|
|
2020-12-21 12:54:26 -07:00
|
|
|
defp hashtags_changed?(_, _), do: false
|
|
|
|
|
2019-05-21 11:40:35 -06:00
|
|
|
def get_by_id(nil), do: nil
|
|
|
|
def get_by_id(id), do: Repo.get(Object, id)
|
|
|
|
|
2019-09-18 09:13:21 -06:00
|
|
|
def get_by_id_and_maybe_refetch(id, opts \\ []) do
|
|
|
|
%{updated_at: updated_at} = object = get_by_id(id)
|
|
|
|
|
|
|
|
if opts[:interval] &&
|
2019-09-18 09:52:33 -06:00
|
|
|
NaiveDateTime.diff(NaiveDateTime.utc_now(), updated_at) > opts[:interval] do
|
2019-09-18 09:13:21 -06:00
|
|
|
case Fetcher.refetch_object(object) do
|
|
|
|
{:ok, %Object{} = object} ->
|
|
|
|
object
|
|
|
|
|
|
|
|
e ->
|
|
|
|
Logger.error("Couldn't refresh #{object.data["id"]}:\n#{inspect(e)}")
|
|
|
|
object
|
|
|
|
end
|
|
|
|
else
|
|
|
|
object
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2017-10-24 00:39:24 -06:00
|
|
|
def get_by_ap_id(nil), do: nil
|
2018-03-30 07:01:53 -06:00
|
|
|
|
2017-03-30 10:07:01 -06:00
|
|
|
def get_by_ap_id(ap_id) do
|
2018-03-30 07:01:53 -06:00
|
|
|
Repo.one(from(object in Object, where: fragment("(?)->>'id' = ?", object.data, ^ap_id)))
|
2017-03-30 10:07:01 -06:00
|
|
|
end
|
2017-04-30 05:53:26 -06:00
|
|
|
|
2020-01-12 11:48:58 -07:00
|
|
|
@doc """
|
|
|
|
Get a single attachment by it's name and href
|
|
|
|
"""
|
|
|
|
@spec get_attachment_by_name_and_href(String.t(), String.t()) :: Object.t() | nil
|
|
|
|
def get_attachment_by_name_and_href(name, href) do
|
|
|
|
query =
|
|
|
|
from(o in Object,
|
|
|
|
where: fragment("(?)->>'name' = ?", o.data, ^name),
|
|
|
|
where: fragment("(?)->>'href' = ?", o.data, ^href)
|
|
|
|
)
|
|
|
|
|
|
|
|
Repo.one(query)
|
|
|
|
end
|
|
|
|
|
2019-07-08 10:53:02 -06:00
|
|
|
defp warn_on_no_object_preloaded(ap_id) do
|
2019-11-21 06:03:01 -07:00
|
|
|
"Object.normalize() called without preloaded object (#{inspect(ap_id)}). Consider preloading the object"
|
2019-07-08 10:53:02 -06:00
|
|
|
|> Logger.debug()
|
|
|
|
|
|
|
|
Logger.debug("Backtrace: #{inspect(Process.info(:erlang.self(), :current_stacktrace))}")
|
|
|
|
end
|
|
|
|
|
2022-09-10 20:54:04 -06:00
|
|
|
def normalize(_, options \\ [fetch: false, id_only: false])
|
2019-07-08 10:53:02 -06:00
|
|
|
|
2019-03-22 17:34:47 -06:00
|
|
|
# If we pass an Activity to Object.normalize(), we can try to use the preloaded object.
|
|
|
|
# Use this whenever possible, especially when walking graphs in an O(N) loop!
|
2021-01-04 05:38:31 -07:00
|
|
|
def normalize(%Object{} = object, _), do: object
|
|
|
|
def normalize(%Activity{object: %Object{} = object}, _), do: object
|
2018-12-03 20:17:25 -07:00
|
|
|
|
2019-04-01 02:55:59 -06:00
|
|
|
# A hack for fake activities
|
2021-01-04 05:38:31 -07:00
|
|
|
def normalize(%Activity{data: %{"object" => %{"fake" => true} = data}}, _) do
|
2019-04-01 02:55:59 -06:00
|
|
|
%Object{id: "pleroma:fake_object_id", data: data}
|
|
|
|
end
|
|
|
|
|
2019-07-08 10:53:02 -06:00
|
|
|
# No preloaded object
|
2021-01-04 05:38:31 -07:00
|
|
|
def normalize(%Activity{data: %{"object" => %{"id" => ap_id}}}, options) do
|
2019-07-08 10:53:02 -06:00
|
|
|
warn_on_no_object_preloaded(ap_id)
|
2021-01-04 05:38:31 -07:00
|
|
|
normalize(ap_id, options)
|
2019-03-22 17:34:47 -06:00
|
|
|
end
|
|
|
|
|
2019-07-08 10:53:02 -06:00
|
|
|
# No preloaded object
|
2021-01-04 05:38:31 -07:00
|
|
|
def normalize(%Activity{data: %{"object" => ap_id}}, options) do
|
2019-07-08 10:53:02 -06:00
|
|
|
warn_on_no_object_preloaded(ap_id)
|
2021-01-04 05:38:31 -07:00
|
|
|
normalize(ap_id, options)
|
2019-03-22 17:34:47 -06:00
|
|
|
end
|
|
|
|
|
|
|
|
# Old way, try fetching the object through cache.
|
2021-01-04 05:38:31 -07:00
|
|
|
def normalize(%{"id" => ap_id}, options), do: normalize(ap_id, options)
|
2019-06-29 11:04:50 -06:00
|
|
|
|
2021-01-04 05:38:31 -07:00
|
|
|
def normalize(ap_id, options) when is_binary(ap_id) do
|
2022-09-10 20:54:04 -06:00
|
|
|
cond do
|
|
|
|
Keyword.get(options, :id_only) ->
|
|
|
|
ap_id
|
|
|
|
|
|
|
|
Keyword.get(options, :fetch) ->
|
|
|
|
Fetcher.fetch_object_from_id!(ap_id, options)
|
|
|
|
|
|
|
|
true ->
|
|
|
|
get_cached_by_ap_id(ap_id)
|
2021-01-04 05:38:31 -07:00
|
|
|
end
|
2019-06-29 11:04:50 -06:00
|
|
|
end
|
|
|
|
|
2021-01-04 05:38:31 -07:00
|
|
|
def normalize(_, _), do: nil
|
2018-06-17 23:23:54 -06:00
|
|
|
|
2020-05-18 00:51:53 -06:00
|
|
|
# Owned objects can only be accessed by their owner
|
|
|
|
def authorize_access(%Object{data: %{"actor" => actor}}, %User{ap_id: ap_id}) do
|
|
|
|
if actor == ap_id do
|
|
|
|
:ok
|
|
|
|
else
|
|
|
|
{:error, :forbidden}
|
|
|
|
end
|
|
|
|
end
|
2018-12-06 00:26:17 -07:00
|
|
|
|
2020-05-18 00:51:53 -06:00
|
|
|
# Legacy objects can be accessed by anybody
|
|
|
|
def authorize_access(%Object{}, %User{}), do: :ok
|
2018-12-06 00:26:17 -07:00
|
|
|
|
2019-12-30 01:30:20 -07:00
|
|
|
@spec get_cached_by_ap_id(String.t()) :: Object.t() | nil
|
2019-02-03 10:28:14 -07:00
|
|
|
def get_cached_by_ap_id(ap_id) do
|
|
|
|
key = "object:#{ap_id}"
|
|
|
|
|
2020-12-18 09:44:46 -07:00
|
|
|
with {:ok, nil} <- @cachex.get(:object_cache, key),
|
2019-12-30 01:30:20 -07:00
|
|
|
object when not is_nil(object) <- get_by_ap_id(ap_id),
|
2020-12-18 09:44:46 -07:00
|
|
|
{:ok, true} <- @cachex.put(:object_cache, key, object) do
|
2019-12-30 01:30:20 -07:00
|
|
|
object
|
|
|
|
else
|
|
|
|
{:ok, object} -> object
|
|
|
|
nil -> nil
|
|
|
|
end
|
2017-05-01 08:12:20 -06:00
|
|
|
end
|
|
|
|
|
2018-12-24 17:00:06 -07:00
|
|
|
def make_tombstone(%Object{data: %{"id" => id, "type" => type}}, deleted \\ DateTime.utc_now()) do
|
|
|
|
%ObjectTombstone{
|
|
|
|
id: id,
|
|
|
|
formerType: type,
|
2018-12-23 16:25:36 -07:00
|
|
|
deleted: deleted
|
|
|
|
}
|
2018-12-24 17:00:06 -07:00
|
|
|
|> Map.from_struct()
|
2018-12-23 16:25:36 -07:00
|
|
|
end
|
|
|
|
|
2018-12-24 17:00:06 -07:00
|
|
|
def swap_object_with_tombstone(object) do
|
|
|
|
tombstone = make_tombstone(object)
|
2018-12-23 16:25:36 -07:00
|
|
|
|
2021-02-11 09:30:21 -07:00
|
|
|
with {:ok, object} <-
|
|
|
|
object
|
|
|
|
|> Object.change(%{data: tombstone})
|
|
|
|
|> Repo.update() do
|
|
|
|
Hashtag.unlink(object)
|
|
|
|
{:ok, object}
|
|
|
|
end
|
2018-12-23 16:25:36 -07:00
|
|
|
end
|
|
|
|
|
2018-11-01 01:28:48 -06:00
|
|
|
def delete(%Object{data: %{"id" => id}} = object) do
|
2018-12-24 17:00:06 -07:00
|
|
|
with {:ok, _obj} = swap_object_with_tombstone(object),
|
2019-12-04 16:50:38 -07:00
|
|
|
deleted_activity = Activity.delete_all_by_object_ap_id(id),
|
2020-05-15 12:34:46 -06:00
|
|
|
{:ok, _} <- invalid_object_cache(object) do
|
|
|
|
cleanup_attachments(
|
|
|
|
Config.get([:instance, :cleanup_attachments]),
|
|
|
|
%{"object" => object}
|
|
|
|
)
|
2020-01-30 15:20:37 -07:00
|
|
|
|
2019-03-09 04:12:15 -07:00
|
|
|
{:ok, object, deleted_activity}
|
2018-11-01 01:28:48 -06:00
|
|
|
end
|
|
|
|
end
|
2019-02-03 10:28:14 -07:00
|
|
|
|
2020-05-15 12:34:46 -06:00
|
|
|
@spec cleanup_attachments(boolean(), %{required(:object) => map()}) ::
|
|
|
|
{:ok, Oban.Job.t() | nil}
|
|
|
|
def cleanup_attachments(true, %{"object" => _} = params) do
|
|
|
|
AttachmentsCleanupWorker.enqueue("cleanup_attachments", params)
|
|
|
|
end
|
|
|
|
|
|
|
|
def cleanup_attachments(_, _), do: {:ok, nil}
|
|
|
|
|
|
|
|
def prune(%Object{data: %{"id" => _id}} = object) do
|
2019-05-20 18:41:40 -06:00
|
|
|
with {:ok, object} <- Repo.delete(object),
|
2020-05-15 12:34:46 -06:00
|
|
|
{:ok, _} <- invalid_object_cache(object) do
|
2019-05-20 18:41:40 -06:00
|
|
|
{:ok, object}
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2020-05-15 12:34:46 -06:00
|
|
|
def invalid_object_cache(%Object{data: %{"id" => id}}) do
|
2020-12-18 09:44:46 -07:00
|
|
|
with {:ok, true} <- @cachex.del(:object_cache, "object:#{id}") do
|
|
|
|
@cachex.del(:web_resp_cache, URI.parse(id).path)
|
2020-05-15 12:34:46 -06:00
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2019-02-03 10:28:14 -07:00
|
|
|
def set_cache(%Object{data: %{"id" => ap_id}} = object) do
|
2020-12-18 09:44:46 -07:00
|
|
|
@cachex.put(:object_cache, "object:#{ap_id}", object)
|
2019-02-03 10:28:14 -07:00
|
|
|
{:ok, object}
|
|
|
|
end
|
|
|
|
|
|
|
|
def update_and_set_cache(changeset) do
|
|
|
|
with {:ok, object} <- Repo.update(changeset) do
|
|
|
|
set_cache(object)
|
|
|
|
end
|
|
|
|
end
|
2019-03-25 11:21:48 -06:00
|
|
|
|
|
|
|
def increase_replies_count(ap_id) do
|
|
|
|
Object
|
|
|
|
|> where([o], fragment("?->>'id' = ?::text", o.data, ^to_string(ap_id)))
|
|
|
|
|> update([o],
|
|
|
|
set: [
|
|
|
|
data:
|
|
|
|
fragment(
|
|
|
|
"""
|
2019-10-18 05:11:30 -06:00
|
|
|
safe_jsonb_set(?, '{repliesCount}',
|
2019-03-25 11:21:48 -06:00
|
|
|
(coalesce((?->>'repliesCount')::int, 0) + 1)::varchar::jsonb, true)
|
|
|
|
""",
|
|
|
|
o.data,
|
|
|
|
o.data
|
|
|
|
)
|
|
|
|
]
|
|
|
|
)
|
|
|
|
|> Repo.update_all([])
|
|
|
|
|> case do
|
|
|
|
{1, [object]} -> set_cache(object)
|
|
|
|
_ -> {:error, "Not found"}
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2020-08-05 07:40:32 -06:00
|
|
|
defp poll_is_multiple?(%Object{data: %{"anyOf" => [_ | _]}}), do: true
|
2020-06-14 16:30:45 -06:00
|
|
|
|
|
|
|
defp poll_is_multiple?(_), do: false
|
|
|
|
|
2019-03-25 11:21:48 -06:00
|
|
|
def decrease_replies_count(ap_id) do
|
|
|
|
Object
|
|
|
|
|> where([o], fragment("?->>'id' = ?::text", o.data, ^to_string(ap_id)))
|
|
|
|
|> update([o],
|
|
|
|
set: [
|
|
|
|
data:
|
|
|
|
fragment(
|
|
|
|
"""
|
2019-10-18 05:11:30 -06:00
|
|
|
safe_jsonb_set(?, '{repliesCount}',
|
2019-03-25 11:21:48 -06:00
|
|
|
(greatest(0, (?->>'repliesCount')::int - 1))::varchar::jsonb, true)
|
|
|
|
""",
|
|
|
|
o.data,
|
|
|
|
o.data
|
|
|
|
)
|
|
|
|
]
|
|
|
|
)
|
|
|
|
|> Repo.update_all([])
|
|
|
|
|> case do
|
|
|
|
{1, [object]} -> set_cache(object)
|
|
|
|
_ -> {:error, "Not found"}
|
2018-11-01 01:28:48 -06:00
|
|
|
end
|
|
|
|
end
|
2019-05-21 05:12:10 -06:00
|
|
|
|
2020-04-22 06:06:39 -06:00
|
|
|
def increase_vote_count(ap_id, name, actor) do
|
2021-01-04 05:38:31 -07:00
|
|
|
with %Object{} = object <- Object.normalize(ap_id, fetch: false),
|
2019-05-21 05:12:10 -06:00
|
|
|
"Question" <- object.data["type"] do
|
2020-06-14 16:30:45 -06:00
|
|
|
key = if poll_is_multiple?(object), do: "anyOf", else: "oneOf"
|
2019-05-21 05:12:10 -06:00
|
|
|
|
|
|
|
options =
|
2020-06-14 16:30:45 -06:00
|
|
|
object.data[key]
|
2019-05-21 05:12:10 -06:00
|
|
|
|> Enum.map(fn
|
|
|
|
%{"name" => ^name} = option ->
|
|
|
|
Kernel.update_in(option["replies"]["totalItems"], &(&1 + 1))
|
|
|
|
|
|
|
|
option ->
|
|
|
|
option
|
|
|
|
end)
|
|
|
|
|
2020-04-22 06:06:39 -06:00
|
|
|
voters = [actor | object.data["voters"] || []] |> Enum.uniq()
|
|
|
|
|
2019-05-21 05:12:10 -06:00
|
|
|
data =
|
2020-06-14 16:30:45 -06:00
|
|
|
object.data
|
|
|
|
|> Map.put(key, options)
|
2020-04-22 06:06:39 -06:00
|
|
|
|> Map.put("voters", voters)
|
2019-05-21 05:12:10 -06:00
|
|
|
|
|
|
|
object
|
|
|
|
|> Object.change(%{data: data})
|
|
|
|
|> update_and_set_cache()
|
|
|
|
else
|
|
|
|
_ -> :noop
|
|
|
|
end
|
|
|
|
end
|
2019-09-06 12:50:00 -06:00
|
|
|
|
|
|
|
@doc "Updates data field of an object"
|
|
|
|
def update_data(%Object{data: data} = object, attrs \\ %{}) do
|
|
|
|
object
|
|
|
|
|> Object.change(%{data: Map.merge(data || %{}, attrs)})
|
|
|
|
|> Repo.update()
|
|
|
|
end
|
2019-11-23 12:55:41 -07:00
|
|
|
|
|
|
|
def local?(%Object{data: %{"id" => id}}) do
|
2021-05-31 14:09:11 -06:00
|
|
|
String.starts_with?(id, Pleroma.Web.Endpoint.url() <> "/")
|
2019-11-23 12:55:41 -07:00
|
|
|
end
|
2020-02-09 00:17:21 -07:00
|
|
|
|
|
|
|
def replies(object, opts \\ []) do
|
2021-01-04 05:38:31 -07:00
|
|
|
object = Object.normalize(object, fetch: false)
|
2020-02-09 00:17:21 -07:00
|
|
|
|
|
|
|
query =
|
|
|
|
Object
|
|
|
|
|> where(
|
|
|
|
[o],
|
|
|
|
fragment("(?)->>'inReplyTo' = ?", o.data, ^object.data["id"])
|
|
|
|
)
|
|
|
|
|> order_by([o], asc: o.id)
|
|
|
|
|
|
|
|
if opts[:self_only] do
|
|
|
|
actor = object.data["actor"]
|
|
|
|
where(query, [o], fragment("(?)->>'actor' = ?", o.data, ^actor))
|
|
|
|
else
|
|
|
|
query
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def self_replies(object, opts \\ []),
|
|
|
|
do: replies(object, Keyword.put(opts, :self_only, true))
|
2020-12-21 12:54:26 -07:00
|
|
|
|
|
|
|
def tags(%Object{data: %{"tag" => tags}}) when is_list(tags), do: tags
|
|
|
|
|
|
|
|
def tags(_), do: []
|
|
|
|
|
2020-12-27 14:08:09 -07:00
|
|
|
def hashtags(%Object{} = object) do
|
2021-01-25 10:12:09 -07:00
|
|
|
# Note: always using embedded hashtags regardless whether they are migrated to hashtags table
|
|
|
|
# (embedded hashtags stay in sync anyways, and we avoid extra joins and preload hassle)
|
|
|
|
embedded_hashtags(object)
|
2020-12-27 14:08:09 -07:00
|
|
|
end
|
2020-12-21 12:54:26 -07:00
|
|
|
|
2021-01-25 10:12:09 -07:00
|
|
|
def embedded_hashtags(%Object{data: data}) do
|
2020-12-21 12:54:26 -07:00
|
|
|
object_data_hashtags(data)
|
|
|
|
end
|
|
|
|
|
2021-01-25 10:12:09 -07:00
|
|
|
def embedded_hashtags(_), do: []
|
2020-12-21 12:54:26 -07:00
|
|
|
|
2021-01-07 02:20:29 -07:00
|
|
|
def object_data_hashtags(%{"tag" => tags}) when is_list(tags) do
|
2020-12-30 04:42:35 -07:00
|
|
|
tags
|
|
|
|
|> Enum.filter(fn
|
|
|
|
%{"type" => "Hashtag"} = data -> Map.has_key?(data, "name")
|
2020-12-30 05:22:49 -07:00
|
|
|
plain_text when is_bitstring(plain_text) -> true
|
2020-12-30 04:42:35 -07:00
|
|
|
_ -> false
|
|
|
|
end)
|
|
|
|
|> Enum.map(fn
|
|
|
|
%{"name" => "#" <> hashtag} -> String.downcase(hashtag)
|
|
|
|
%{"name" => hashtag} -> String.downcase(hashtag)
|
2020-12-30 05:22:49 -07:00
|
|
|
hashtag when is_bitstring(hashtag) -> String.downcase(hashtag)
|
2020-12-30 04:42:35 -07:00
|
|
|
end)
|
2020-12-30 05:22:49 -07:00
|
|
|
|> Enum.uniq()
|
2021-01-21 10:50:06 -07:00
|
|
|
# Note: "" elements (plain text) might occur in `data.tag` for incoming objects
|
|
|
|
|> Enum.filter(&(&1 not in [nil, ""]))
|
2020-12-21 12:54:26 -07:00
|
|
|
end
|
|
|
|
|
2021-01-07 02:20:29 -07:00
|
|
|
def object_data_hashtags(_), do: []
|
2017-03-21 02:21:52 -06:00
|
|
|
end
|