2018-12-23 12:04:54 -08:00
|
|
|
# Pleroma: A lightweight social networking server
|
2020-03-01 21:08:45 -08:00
|
|
|
# Copyright © 2017-2020 Pleroma Authors <https://pleroma.social/>
|
2018-12-23 12:04:54 -08:00
|
|
|
# SPDX-License-Identifier: AGPL-3.0-only
|
|
|
|
|
2018-09-09 16:29:00 -07:00
|
|
|
defmodule Pleroma.HTML do
|
2019-12-08 08:42:40 -08:00
|
|
|
# Scrubbers are compiled on boot so they can be configured in OTP releases
|
|
|
|
# @on_load :compile_scrubbers
|
|
|
|
|
|
|
|
def compile_scrubbers do
|
|
|
|
dir = Path.join(:code.priv_dir(:pleroma), "scrubbers")
|
|
|
|
|
|
|
|
dir
|
2019-12-09 09:38:01 -08:00
|
|
|
|> Pleroma.Utils.compile_dir()
|
2019-12-08 08:42:40 -08:00
|
|
|
|> case do
|
|
|
|
{:error, _errors, _warnings} ->
|
|
|
|
raise "Compiling scrubbers failed"
|
|
|
|
|
|
|
|
{:ok, _modules, _warnings} ->
|
|
|
|
:ok
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2018-09-15 19:07:01 -07:00
|
|
|
defp get_scrubbers(scrubber) when is_atom(scrubber), do: [scrubber]
|
|
|
|
defp get_scrubbers(scrubbers) when is_list(scrubbers), do: scrubbers
|
|
|
|
defp get_scrubbers(_), do: [Pleroma.HTML.Scrubber.Default]
|
|
|
|
|
2019-03-04 19:18:43 -08:00
|
|
|
def get_scrubbers do
|
2018-11-06 10:34:57 -08:00
|
|
|
Pleroma.Config.get([:markup, :scrub_policy])
|
2018-09-15 19:07:01 -07:00
|
|
|
|> get_scrubbers
|
|
|
|
end
|
|
|
|
|
2018-09-21 18:10:53 -07:00
|
|
|
def filter_tags(html, nil) do
|
2018-12-30 11:44:17 -08:00
|
|
|
filter_tags(html, get_scrubbers())
|
|
|
|
end
|
|
|
|
|
|
|
|
def filter_tags(html, scrubbers) when is_list(scrubbers) do
|
|
|
|
Enum.reduce(scrubbers, html, fn scrubber, html ->
|
2018-09-15 19:07:01 -07:00
|
|
|
filter_tags(html, scrubber)
|
|
|
|
end)
|
2018-09-09 16:29:00 -07:00
|
|
|
end
|
|
|
|
|
2019-10-28 15:18:08 -07:00
|
|
|
def filter_tags(html, scrubber) do
|
|
|
|
{:ok, content} = FastSanitize.Sanitizer.scrub(html, scrubber)
|
|
|
|
content
|
|
|
|
end
|
|
|
|
|
2018-09-21 18:10:53 -07:00
|
|
|
def filter_tags(html), do: filter_tags(html, nil)
|
2019-10-28 15:18:08 -07:00
|
|
|
def strip_tags(html), do: filter_tags(html, FastSanitize.Sanitizer.StripTags)
|
2018-12-30 23:19:48 -08:00
|
|
|
|
2019-04-30 12:52:17 -07:00
|
|
|
def get_cached_scrubbed_html_for_activity(
|
|
|
|
content,
|
|
|
|
scrubbers,
|
|
|
|
activity,
|
|
|
|
key \\ "",
|
|
|
|
callback \\ fn x -> x end
|
|
|
|
) do
|
2019-04-05 05:19:44 -07:00
|
|
|
key = "#{key}#{generate_scrubber_signature(scrubbers)}|#{activity.id}"
|
2019-10-29 10:58:54 -07:00
|
|
|
|
2019-04-01 01:55:59 -07:00
|
|
|
Cachex.fetch!(:scrubber_cache, key, fn _key ->
|
2019-04-17 02:22:32 -07:00
|
|
|
object = Pleroma.Object.normalize(activity)
|
2019-04-30 12:52:17 -07:00
|
|
|
ensure_scrubbed_html(content, scrubbers, object.data["fake"] || false, callback)
|
2019-04-01 01:55:59 -07:00
|
|
|
end)
|
2018-12-30 23:19:48 -08:00
|
|
|
end
|
|
|
|
|
2019-04-05 05:19:44 -07:00
|
|
|
def get_cached_stripped_html_for_activity(content, activity, key) do
|
|
|
|
get_cached_scrubbed_html_for_activity(
|
2019-01-04 15:25:31 -08:00
|
|
|
content,
|
2019-10-28 15:18:08 -07:00
|
|
|
FastSanitize.Sanitizer.StripTags,
|
2019-04-05 05:19:44 -07:00
|
|
|
activity,
|
2019-04-30 12:52:17 -07:00
|
|
|
key,
|
|
|
|
&HtmlEntities.decode/1
|
2019-01-04 15:25:31 -08:00
|
|
|
)
|
2018-09-21 18:10:53 -07:00
|
|
|
end
|
|
|
|
|
2018-12-30 23:19:48 -08:00
|
|
|
def ensure_scrubbed_html(
|
|
|
|
content,
|
2019-04-01 01:55:59 -07:00
|
|
|
scrubbers,
|
2019-04-30 12:52:17 -07:00
|
|
|
fake,
|
|
|
|
callback
|
2018-12-30 23:19:48 -08:00
|
|
|
) do
|
2019-04-30 12:52:17 -07:00
|
|
|
content =
|
|
|
|
content
|
|
|
|
|> filter_tags(scrubbers)
|
|
|
|
|> callback.()
|
|
|
|
|
|
|
|
if fake do
|
|
|
|
{:ignore, content}
|
|
|
|
else
|
|
|
|
{:commit, content}
|
|
|
|
end
|
2018-12-30 23:19:48 -08:00
|
|
|
end
|
|
|
|
|
|
|
|
defp generate_scrubber_signature(scrubber) when is_atom(scrubber) do
|
|
|
|
generate_scrubber_signature([scrubber])
|
|
|
|
end
|
|
|
|
|
|
|
|
defp generate_scrubber_signature(scrubbers) do
|
|
|
|
Enum.reduce(scrubbers, "", fn scrubber, signature ->
|
2019-01-04 15:19:46 -08:00
|
|
|
"#{signature}#{to_string(scrubber)}"
|
2018-12-30 23:19:48 -08:00
|
|
|
end)
|
|
|
|
end
|
2018-09-21 18:10:53 -07:00
|
|
|
|
2019-02-04 21:06:17 -08:00
|
|
|
def extract_first_external_url(_, nil), do: {:error, "No content"}
|
|
|
|
|
2019-01-26 06:55:12 -08:00
|
|
|
def extract_first_external_url(object, content) do
|
|
|
|
key = "URL|#{object.id}"
|
|
|
|
|
|
|
|
Cachex.fetch!(:scrubber_cache, key, fn _key ->
|
|
|
|
result =
|
|
|
|
content
|
2020-02-15 14:55:26 -08:00
|
|
|
|> Floki.parse_fragment!()
|
2019-06-18 14:31:30 -07:00
|
|
|
|> Floki.filter_out("a.mention,a.hashtag,a[rel~=\"tag\"]")
|
2019-01-26 06:55:12 -08:00
|
|
|
|> Floki.attribute("a", "href")
|
|
|
|
|> Enum.at(0)
|
|
|
|
|
2019-01-28 12:55:33 -08:00
|
|
|
{:commit, {:ok, result}}
|
2019-01-26 06:55:12 -08:00
|
|
|
end)
|
2018-09-09 16:29:00 -07:00
|
|
|
end
|
|
|
|
end
|