2018-12-23 13:04:54 -07:00
|
|
|
# Pleroma: A lightweight social networking server
|
2018-12-31 08:41:47 -07:00
|
|
|
# Copyright © 2017-2019 Pleroma Authors <https://pleroma.social/>
|
2018-12-23 13:04:54 -07:00
|
|
|
# SPDX-License-Identifier: AGPL-3.0-only
|
|
|
|
|
2017-03-21 02:21:52 -06:00
|
|
|
defmodule Pleroma.Object do
|
|
|
|
use Ecto.Schema
|
2019-02-09 08:16:26 -07:00
|
|
|
|
|
|
|
alias Pleroma.Activity
|
2019-03-04 19:52:23 -07:00
|
|
|
alias Pleroma.Object
|
2019-02-09 08:16:26 -07:00
|
|
|
alias Pleroma.ObjectTombstone
|
2019-03-04 19:52:23 -07:00
|
|
|
alias Pleroma.Repo
|
|
|
|
alias Pleroma.User
|
2019-02-09 08:16:26 -07:00
|
|
|
|
|
|
|
import Ecto.Query
|
|
|
|
import Ecto.Changeset
|
2017-03-21 02:21:52 -06:00
|
|
|
|
2019-03-22 17:34:47 -06:00
|
|
|
require Logger
|
|
|
|
|
2017-03-21 02:21:52 -06:00
|
|
|
schema "objects" do
|
2018-03-30 07:01:53 -06:00
|
|
|
field(:data, :map)
|
2017-03-21 02:21:52 -06:00
|
|
|
|
|
|
|
timestamps()
|
|
|
|
end
|
2017-03-30 10:07:01 -06:00
|
|
|
|
2017-05-16 07:31:11 -06:00
|
|
|
def create(data) do
|
|
|
|
Object.change(%Object{}, %{data: data})
|
2019-02-11 01:07:39 -07:00
|
|
|
|> Repo.insert()
|
2017-05-16 07:31:11 -06:00
|
|
|
end
|
|
|
|
|
2017-05-09 10:11:51 -06:00
|
|
|
def change(struct, params \\ %{}) do
|
2017-11-18 18:22:07 -07:00
|
|
|
struct
|
2017-05-09 10:11:51 -06:00
|
|
|
|> cast(params, [:data])
|
|
|
|
|> validate_required([:data])
|
|
|
|
|> unique_constraint(:ap_id, name: :objects_unique_apid_index)
|
|
|
|
end
|
|
|
|
|
2017-10-24 00:39:24 -06:00
|
|
|
def get_by_ap_id(nil), do: nil
|
2018-03-30 07:01:53 -06:00
|
|
|
|
2017-03-30 10:07:01 -06:00
|
|
|
def get_by_ap_id(ap_id) do
|
2018-03-30 07:01:53 -06:00
|
|
|
Repo.one(from(object in Object, where: fragment("(?)->>'id' = ?", object.data, ^ap_id)))
|
2017-03-30 10:07:01 -06:00
|
|
|
end
|
2017-04-30 05:53:26 -06:00
|
|
|
|
2019-03-22 17:34:47 -06:00
|
|
|
# If we pass an Activity to Object.normalize(), we can try to use the preloaded object.
|
|
|
|
# Use this whenever possible, especially when walking graphs in an O(N) loop!
|
|
|
|
def normalize(%Activity{object: %Object{} = object}), do: object
|
|
|
|
|
|
|
|
# Catch and log Object.normalize() calls where the Activity's child object is not
|
|
|
|
# preloaded.
|
|
|
|
def normalize(%Activity{data: %{"object" => %{"id" => ap_id}}}) do
|
2019-03-22 21:00:04 -06:00
|
|
|
Logger.debug(
|
2019-03-22 17:34:47 -06:00
|
|
|
"Object.normalize() called without preloaded object (#{ap_id}). Consider preloading the object!"
|
|
|
|
)
|
|
|
|
|
2019-03-22 21:00:04 -06:00
|
|
|
Logger.debug("Backtrace: #{inspect(Process.info(:erlang.self(), :current_stacktrace))}")
|
|
|
|
|
2019-03-22 17:34:47 -06:00
|
|
|
normalize(ap_id)
|
|
|
|
end
|
|
|
|
|
|
|
|
def normalize(%Activity{data: %{"object" => ap_id}}) do
|
2019-03-22 21:00:04 -06:00
|
|
|
Logger.debug(
|
2019-03-22 17:34:47 -06:00
|
|
|
"Object.normalize() called without preloaded object (#{ap_id}). Consider preloading the object!"
|
|
|
|
)
|
|
|
|
|
2019-03-22 21:00:04 -06:00
|
|
|
Logger.debug("Backtrace: #{inspect(Process.info(:erlang.self(), :current_stacktrace))}")
|
|
|
|
|
2019-03-22 17:34:47 -06:00
|
|
|
normalize(ap_id)
|
|
|
|
end
|
|
|
|
|
|
|
|
# Old way, try fetching the object through cache.
|
2019-02-01 12:38:57 -07:00
|
|
|
def normalize(%{"id" => ap_id}), do: normalize(ap_id)
|
|
|
|
def normalize(ap_id) when is_binary(ap_id), do: get_cached_by_ap_id(ap_id)
|
2018-06-17 23:23:54 -06:00
|
|
|
def normalize(_), do: nil
|
|
|
|
|
2018-12-06 00:26:17 -07:00
|
|
|
# Owned objects can only be mutated by their owner
|
|
|
|
def authorize_mutation(%Object{data: %{"actor" => actor}}, %User{ap_id: ap_id}),
|
|
|
|
do: actor == ap_id
|
|
|
|
|
|
|
|
# Legacy objects can be mutated by anybody
|
|
|
|
def authorize_mutation(%Object{}, %User{}), do: true
|
|
|
|
|
2019-02-03 10:28:14 -07:00
|
|
|
def get_cached_by_ap_id(ap_id) do
|
|
|
|
key = "object:#{ap_id}"
|
|
|
|
|
|
|
|
Cachex.fetch!(:object_cache, key, fn _ ->
|
|
|
|
object = get_by_ap_id(ap_id)
|
|
|
|
|
|
|
|
if object do
|
|
|
|
{:commit, object}
|
|
|
|
else
|
|
|
|
{:ignore, object}
|
|
|
|
end
|
|
|
|
end)
|
2017-05-01 08:12:20 -06:00
|
|
|
end
|
|
|
|
|
2017-04-30 05:53:26 -06:00
|
|
|
def context_mapping(context) do
|
2017-06-20 01:50:22 -06:00
|
|
|
Object.change(%Object{}, %{data: %{"id" => context}})
|
2017-04-30 05:53:26 -06:00
|
|
|
end
|
2018-11-01 01:28:48 -06:00
|
|
|
|
2018-12-24 17:00:06 -07:00
|
|
|
def make_tombstone(%Object{data: %{"id" => id, "type" => type}}, deleted \\ DateTime.utc_now()) do
|
|
|
|
%ObjectTombstone{
|
|
|
|
id: id,
|
|
|
|
formerType: type,
|
2018-12-23 16:25:36 -07:00
|
|
|
deleted: deleted
|
|
|
|
}
|
2018-12-24 17:00:06 -07:00
|
|
|
|> Map.from_struct()
|
2018-12-23 16:25:36 -07:00
|
|
|
end
|
|
|
|
|
2018-12-24 17:00:06 -07:00
|
|
|
def swap_object_with_tombstone(object) do
|
|
|
|
tombstone = make_tombstone(object)
|
2018-12-23 16:25:36 -07:00
|
|
|
|
|
|
|
object
|
|
|
|
|> Object.change(%{data: tombstone})
|
|
|
|
|> Repo.update()
|
|
|
|
end
|
|
|
|
|
2018-11-01 01:28:48 -06:00
|
|
|
def delete(%Object{data: %{"id" => id}} = object) do
|
2018-12-24 17:00:06 -07:00
|
|
|
with {:ok, _obj} = swap_object_with_tombstone(object),
|
2019-03-09 04:12:15 -07:00
|
|
|
deleted_activity = Activity.delete_by_ap_id(id),
|
2018-11-01 02:30:10 -06:00
|
|
|
{:ok, true} <- Cachex.del(:object_cache, "object:#{id}") do
|
2019-03-09 04:12:15 -07:00
|
|
|
{:ok, object, deleted_activity}
|
2018-11-01 01:28:48 -06:00
|
|
|
end
|
|
|
|
end
|
2019-02-03 10:28:14 -07:00
|
|
|
|
|
|
|
def set_cache(%Object{data: %{"id" => ap_id}} = object) do
|
|
|
|
Cachex.put(:object_cache, "object:#{ap_id}", object)
|
|
|
|
{:ok, object}
|
|
|
|
end
|
|
|
|
|
|
|
|
def update_and_set_cache(changeset) do
|
|
|
|
with {:ok, object} <- Repo.update(changeset) do
|
|
|
|
set_cache(object)
|
|
|
|
else
|
|
|
|
e -> e
|
|
|
|
end
|
|
|
|
end
|
2017-03-21 02:21:52 -06:00
|
|
|
end
|