lemmy/crates/routes/src/feeds.rs

581 lines
15 KiB
Rust
Raw Normal View History

use crate::local_user_view_from_jwt;
use actix_web::{error::ErrorBadRequest, web, Error, HttpRequest, HttpResponse, Result};
use anyhow::anyhow;
use chrono::{DateTime, Utc};
use lemmy_api_common::{context::LemmyContext, utils::check_private_instance};
use lemmy_db_schema::{
source::{community::Community, person::Person},
traits::ApubActor,
CommentSortType,
CommunityVisibility,
ListingType,
SortType,
};
use lemmy_db_views::{
post_view::PostQuery,
structs::{PostView, SiteView},
};
use lemmy_db_views_actor::{
comment_reply_view::CommentReplyQuery,
person_mention_view::PersonMentionQuery,
structs::{CommentReplyView, PersonMentionView},
};
use lemmy_utils::{
cache_header::cache_1hour,
error::{LemmyError, LemmyErrorType},
utils::markdown::{markdown_to_html, sanitize_html},
};
2021-11-22 18:58:31 +00:00
use once_cell::sync::Lazy;
use rss::{
extension::{dublincore::DublinCoreExtension, ExtensionBuilder, ExtensionMap},
Channel,
EnclosureBuilder,
Guid,
Item,
};
2020-05-16 14:04:08 +00:00
use serde::Deserialize;
2021-12-14 13:30:37 +00:00
use std::{collections::BTreeMap, str::FromStr};
2019-11-19 17:07:10 +00:00
const RSS_FETCH_LIMIT: i64 = 20;
#[derive(Deserialize)]
struct Params {
sort: Option<String>,
limit: Option<i64>,
page: Option<i64>,
}
impl Params {
fn sort_type(&self) -> Result<SortType, Error> {
let sort_query = self
.sort
.clone()
.unwrap_or_else(|| SortType::Hot.to_string());
SortType::from_str(&sort_query).map_err(ErrorBadRequest)
}
fn get_limit(&self) -> i64 {
self.limit.unwrap_or(RSS_FETCH_LIMIT)
}
fn get_page(&self) -> i64 {
self.page.unwrap_or(1)
}
}
2019-12-01 19:01:38 +00:00
enum RequestType {
Community,
User,
Front,
Inbox,
2019-12-01 19:01:38 +00:00
}
pub fn config(cfg: &mut web::ServiceConfig) {
cfg.service(
web::scope("/feeds")
.route("/{type}/{name}.xml", web::get().to(get_feed))
.route("/all.xml", web::get().to(get_all_feed).wrap(cache_1hour()))
.route(
"/local.xml",
web::get().to(get_local_feed).wrap(cache_1hour()),
),
);
}
2021-12-14 13:30:37 +00:00
static RSS_NAMESPACE: Lazy<BTreeMap<String, String>> = Lazy::new(|| {
let mut h = BTreeMap::new();
2021-11-22 18:58:31 +00:00
h.insert(
"dc".to_string(),
rss::extension::dublincore::NAMESPACE.to_string(),
);
h.insert(
"media".to_string(),
"http://search.yahoo.com/mrss/".to_string(),
);
2021-11-22 18:58:31 +00:00
h
});
2020-11-20 14:11:47 +00:00
/// Removes any characters disallowed by the XML grammar.
/// See https://www.w3.org/TR/xml/#NT-Char for details.
fn sanitize_xml(input: String) -> String {
input
.chars()
.filter(|&c| {
matches!(c,
'\u{09}'
| '\u{0A}'
| '\u{0D}'
| '\u{20}'..='\u{D7FF}'
| '\u{E000}'..='\u{FFFD}'
| '\u{10000}'..='\u{10FFFF}')
})
.collect()
}
#[tracing::instrument(skip_all)]
async fn get_all_feed(
info: web::Query<Params>,
context: web::Data<LemmyContext>,
) -> Result<HttpResponse, Error> {
Ok(
get_feed_data(
&context,
ListingType::All,
info.sort_type()?,
info.get_limit(),
info.get_page(),
)
.await?,
)
2019-12-01 19:01:38 +00:00
}
#[tracing::instrument(skip_all)]
async fn get_local_feed(
info: web::Query<Params>,
context: web::Data<LemmyContext>,
) -> Result<HttpResponse, Error> {
Ok(
get_feed_data(
&context,
ListingType::Local,
info.sort_type()?,
info.get_limit(),
info.get_page(),
)
.await?,
)
}
#[tracing::instrument(skip_all)]
async fn get_feed_data(
context: &LemmyContext,
listing_type: ListingType,
sort_type: SortType,
limit: i64,
page: i64,
) -> Result<HttpResponse, LemmyError> {
Make functions work with both connection and pool (#3420) * a lot * merge * Fix stuff broken by merge * Get rid of repetitive `&mut *context.conn().await?` * Add blank lines under each line with `conn =` * Fix style mistakes (partial) * Revert "Fix style mistakes (partial)" This reverts commit 48a033b87f4fdc1ce14ff86cc019e1c703cd2741. * Revert "Add blank lines under each line with `conn =`" This reverts commit 773a6d3beba2cf89eac75913078b40c4f5190dd4. * Revert "Get rid of repetitive `&mut *context.conn().await?`" This reverts commit d2c6263ea13710177d49b2791278db5ad115fca5. * Use DbConn for CaptchaAnswer methods * DbConn trait * Remove more `&mut *` * Fix stuff * Re-run CI * try to make ci start * fix * fix * Fix api_common::utils * Fix apub::activities::block * Fix apub::api::resolve_object * Fix some things * Revert "Fix some things" This reverts commit 2bf8574bc8333d8d34ca542d61a0a5b50039c24d. * Revert "Fix apub::api::resolve_object" This reverts commit 3e4059aabbe485b2ff060bdeced8ef958ff62832. * Revert "Fix apub::activities::block" This reverts commit 3b02389abd780a7b1b8a2c89e26febdaa6a12159. * Revert "Fix api_common::utils" This reverts commit 7dc73de613a5618fa57eb06450f3699bbcb41254. * Revert "Revert "Fix api_common::utils"" This reverts commit f740f115e5457e83e53cc223e48196a2c47a9975. * Revert "Revert "Fix apub::activities::block"" This reverts commit 2ee206af7c885c10092cf209bf4a5b1d60327866. * Revert "Revert "Fix apub::api::resolve_object"" This reverts commit 96ed8bf2e9dcadae760743929498312334e23d2e. * Fix fetch_local_site_data * Fix get_comment_parent_creator * Remove unused perma deleted text * Fix routes::feeds * Fix lib.rs * Update lib.rs * rerun ci * Attempt to create custom GetConn and RunQueryDsl traits * Start over * Add GetConn trait * aaaa * Revert "aaaa" This reverts commit acc9ca1aed10c39efdd91cefece066e035a1fe80. * Revert "Revert "aaaa"" This reverts commit 443a2a00a56d152bb7eb429efd0d29a78e21b163. * still aaaaaaaaaaaaa * Return to earlier thing Revert "Add GetConn trait" This reverts commit ab4e94aea5bd9d34cbcddf017339131047e75344. * Try to use DbPool enum * Revert "Try to use DbPool enum" This reverts commit e4d1712646a52006b865a1fbe0dcf79976fdb027. * DbConn and DbPool enums (db_schema only fails to compile for tests) * fmt * Make functions take `&mut DbPool<'_>` and make db_schema tests compile * Add try_join_with_pool macro and run fix-clippy on more crates * Fix some errors * I did it * Remove function variants that take connection * rerun ci * rerun ci * rerun ci
2023-07-11 13:09:59 +00:00
let site_view = SiteView::read_local(&mut context.pool()).await?;
2022-11-09 10:05:00 +00:00
check_private_instance(&None, &site_view.local_site)?;
let posts = PostQuery {
listing_type: (Some(listing_type)),
sort: (Some(sort_type)),
limit: (Some(limit)),
page: (Some(page)),
..Default::default()
}
.list(&site_view.site, &mut context.pool())
.await?;
let items = create_post_items(posts, &context.settings().get_protocol_and_hostname())?;
let mut channel = Channel {
namespaces: RSS_NAMESPACE.clone(),
title: format!("{} - {}", site_view.site.name, listing_type),
link: context.settings().get_protocol_and_hostname(),
items,
..Default::default()
};
2020-12-02 19:32:47 +00:00
if let Some(site_desc) = site_view.site.description {
channel.set_description(&site_desc);
}
let rss = channel.to_string();
Ok(
HttpResponse::Ok()
.content_type("application/rss+xml")
.body(rss),
)
}
#[tracing::instrument(skip_all)]
2020-01-11 12:30:45 +00:00
async fn get_feed(
req: HttpRequest,
2020-01-11 12:30:45 +00:00
info: web::Query<Params>,
context: web::Data<LemmyContext>,
2020-04-21 20:40:03 +00:00
) -> Result<HttpResponse, Error> {
let req_type: String = req.match_info().get("type").unwrap_or("none").parse()?;
let param: String = req.match_info().get("name").unwrap_or("none").parse()?;
2020-09-12 01:37:25 +00:00
let request_type = match req_type.as_str() {
"u" => RequestType::User,
"c" => RequestType::Community,
"front" => RequestType::Front,
"inbox" => RequestType::Inbox,
_ => return Err(ErrorBadRequest(LemmyError::from(anyhow!("wrong_type")))),
};
2022-11-09 10:05:00 +00:00
let builder = match request_type {
RequestType::User => {
get_feed_user(
&context,
&info.sort_type()?,
&info.get_limit(),
&info.get_page(),
&param,
)
.await
2022-11-09 10:05:00 +00:00
}
RequestType::Community => {
get_feed_community(
&context,
&info.sort_type()?,
&info.get_limit(),
&info.get_page(),
&param,
)
.await
2022-11-09 10:05:00 +00:00
}
RequestType::Front => {
get_feed_front(
&context,
&info.sort_type()?,
&info.get_limit(),
&info.get_page(),
2022-11-09 10:05:00 +00:00
&param,
Make functions work with both connection and pool (#3420) * a lot * merge * Fix stuff broken by merge * Get rid of repetitive `&mut *context.conn().await?` * Add blank lines under each line with `conn =` * Fix style mistakes (partial) * Revert "Fix style mistakes (partial)" This reverts commit 48a033b87f4fdc1ce14ff86cc019e1c703cd2741. * Revert "Add blank lines under each line with `conn =`" This reverts commit 773a6d3beba2cf89eac75913078b40c4f5190dd4. * Revert "Get rid of repetitive `&mut *context.conn().await?`" This reverts commit d2c6263ea13710177d49b2791278db5ad115fca5. * Use DbConn for CaptchaAnswer methods * DbConn trait * Remove more `&mut *` * Fix stuff * Re-run CI * try to make ci start * fix * fix * Fix api_common::utils * Fix apub::activities::block * Fix apub::api::resolve_object * Fix some things * Revert "Fix some things" This reverts commit 2bf8574bc8333d8d34ca542d61a0a5b50039c24d. * Revert "Fix apub::api::resolve_object" This reverts commit 3e4059aabbe485b2ff060bdeced8ef958ff62832. * Revert "Fix apub::activities::block" This reverts commit 3b02389abd780a7b1b8a2c89e26febdaa6a12159. * Revert "Fix api_common::utils" This reverts commit 7dc73de613a5618fa57eb06450f3699bbcb41254. * Revert "Revert "Fix api_common::utils"" This reverts commit f740f115e5457e83e53cc223e48196a2c47a9975. * Revert "Revert "Fix apub::activities::block"" This reverts commit 2ee206af7c885c10092cf209bf4a5b1d60327866. * Revert "Revert "Fix apub::api::resolve_object"" This reverts commit 96ed8bf2e9dcadae760743929498312334e23d2e. * Fix fetch_local_site_data * Fix get_comment_parent_creator * Remove unused perma deleted text * Fix routes::feeds * Fix lib.rs * Update lib.rs * rerun ci * Attempt to create custom GetConn and RunQueryDsl traits * Start over * Add GetConn trait * aaaa * Revert "aaaa" This reverts commit acc9ca1aed10c39efdd91cefece066e035a1fe80. * Revert "Revert "aaaa"" This reverts commit 443a2a00a56d152bb7eb429efd0d29a78e21b163. * still aaaaaaaaaaaaa * Return to earlier thing Revert "Add GetConn trait" This reverts commit ab4e94aea5bd9d34cbcddf017339131047e75344. * Try to use DbPool enum * Revert "Try to use DbPool enum" This reverts commit e4d1712646a52006b865a1fbe0dcf79976fdb027. * DbConn and DbPool enums (db_schema only fails to compile for tests) * fmt * Make functions take `&mut DbPool<'_>` and make db_schema tests compile * Add try_join_with_pool macro and run fix-clippy on more crates * Fix some errors * I did it * Remove function variants that take connection * rerun ci * rerun ci * rerun ci
2023-07-11 13:09:59 +00:00
)
.await
2022-11-09 10:05:00 +00:00
}
RequestType::Inbox => get_feed_inbox(&context, &param).await,
2022-11-09 10:05:00 +00:00
}
.map_err(ErrorBadRequest)?;
let rss = builder.to_string();
Ok(
HttpResponse::Ok()
2019-11-19 17:07:10 +00:00
.content_type("application/rss+xml")
.body(rss),
)
2019-11-19 17:07:10 +00:00
}
2019-11-16 02:17:42 +00:00
#[tracing::instrument(skip_all)]
2022-11-09 10:05:00 +00:00
async fn get_feed_user(
context: &LemmyContext,
sort_type: &SortType,
limit: &i64,
page: &i64,
user_name: &str,
) -> Result<Channel, LemmyError> {
let site_view = SiteView::read_local(&mut context.pool()).await?;
let person = Person::read_from_name(&mut context.pool(), user_name, false).await?;
check_private_instance(&None, &site_view.local_site)?;
let posts = PostQuery {
listing_type: (Some(ListingType::All)),
sort: (Some(*sort_type)),
creator_id: (Some(person.id)),
limit: (Some(*limit)),
page: (Some(*page)),
..Default::default()
}
.list(&site_view.site, &mut context.pool())
.await?;
let items = create_post_items(posts, &context.settings().get_protocol_and_hostname())?;
let channel = Channel {
namespaces: RSS_NAMESPACE.clone(),
title: format!("{} - {}", sanitize_xml(site_view.site.name), person.name),
link: person.actor_id.to_string(),
items,
..Default::default()
};
Ok(channel)
}
#[tracing::instrument(skip_all)]
2022-11-09 10:05:00 +00:00
async fn get_feed_community(
context: &LemmyContext,
sort_type: &SortType,
limit: &i64,
page: &i64,
community_name: &str,
) -> Result<Channel, LemmyError> {
let site_view = SiteView::read_local(&mut context.pool()).await?;
let community = Community::read_from_name(&mut context.pool(), community_name, false).await?;
if community.visibility != CommunityVisibility::Public {
return Err(LemmyErrorType::CouldntFindCommunity.into());
}
2019-11-19 17:07:10 +00:00
check_private_instance(&None, &site_view.local_site)?;
let posts = PostQuery {
sort: (Some(*sort_type)),
community_id: (Some(community.id)),
limit: (Some(*limit)),
page: (Some(*page)),
..Default::default()
}
.list(&site_view.site, &mut context.pool())
.await?;
2019-11-16 02:17:42 +00:00
let items = create_post_items(posts, &context.settings().get_protocol_and_hostname())?;
let mut channel = Channel {
namespaces: RSS_NAMESPACE.clone(),
title: format!("{} - {}", sanitize_xml(site_view.site.name), community.name),
link: community.actor_id.to_string(),
items,
..Default::default()
};
if let Some(community_desc) = community.description {
channel.set_description(markdown_to_html(&community_desc));
}
Ok(channel)
}
#[tracing::instrument(skip_all)]
2022-11-09 10:05:00 +00:00
async fn get_feed_front(
context: &LemmyContext,
sort_type: &SortType,
limit: &i64,
page: &i64,
jwt: &str,
) -> Result<Channel, LemmyError> {
let site_view = SiteView::read_local(&mut context.pool()).await?;
let local_user = local_user_view_from_jwt(jwt, context).await?;
check_private_instance(&Some(local_user.clone()), &site_view.local_site)?;
let posts = PostQuery {
listing_type: (Some(ListingType::Subscribed)),
local_user: (Some(&local_user)),
sort: (Some(*sort_type)),
limit: (Some(*limit)),
page: (Some(*page)),
..Default::default()
}
.list(&site_view.site, &mut context.pool())
.await?;
let protocol_and_hostname = context.settings().get_protocol_and_hostname();
let items = create_post_items(posts, &protocol_and_hostname)?;
let mut channel = Channel {
namespaces: RSS_NAMESPACE.clone(),
title: format!("{} - Subscribed", sanitize_xml(site_view.site.name)),
link: protocol_and_hostname,
items,
..Default::default()
};
2020-12-02 19:32:47 +00:00
if let Some(site_desc) = site_view.site.description {
channel.set_description(markdown_to_html(&site_desc));
}
Ok(channel)
}
#[tracing::instrument(skip_all)]
async fn get_feed_inbox(context: &LemmyContext, jwt: &str) -> Result<Channel, LemmyError> {
let site_view = SiteView::read_local(&mut context.pool()).await?;
let local_user = local_user_view_from_jwt(jwt, context).await?;
let person_id = local_user.local_user.person_id;
let show_bot_accounts = local_user.local_user.show_bot_accounts;
let sort = CommentSortType::New;
check_private_instance(&Some(local_user.clone()), &site_view.local_site)?;
let replies = CommentReplyQuery {
recipient_id: (Some(person_id)),
my_person_id: (Some(person_id)),
show_bot_accounts: (show_bot_accounts),
sort: (Some(sort)),
limit: (Some(RSS_FETCH_LIMIT)),
..Default::default()
}
.list(&mut context.pool())
.await?;
let mentions = PersonMentionQuery {
recipient_id: (Some(person_id)),
my_person_id: (Some(person_id)),
show_bot_accounts: (show_bot_accounts),
sort: (Some(sort)),
limit: (Some(RSS_FETCH_LIMIT)),
..Default::default()
}
.list(&mut context.pool())
.await?;
let protocol_and_hostname = context.settings().get_protocol_and_hostname();
let items = create_reply_and_mention_items(replies, mentions, &protocol_and_hostname)?;
let mut channel = Channel {
namespaces: RSS_NAMESPACE.clone(),
title: format!("{} - Inbox", sanitize_xml(site_view.site.name)),
link: format!("{protocol_and_hostname}/inbox"),
items,
..Default::default()
};
2020-12-02 19:32:47 +00:00
if let Some(site_desc) = site_view.site.description {
channel.set_description(&site_desc);
}
Ok(channel)
}
#[tracing::instrument(skip_all)]
fn create_reply_and_mention_items(
replies: Vec<CommentReplyView>,
2021-03-10 22:33:55 +00:00
mentions: Vec<PersonMentionView>,
protocol_and_hostname: &str,
) -> Result<Vec<Item>, LemmyError> {
let mut reply_items: Vec<Item> = replies
.iter()
.map(|r| {
let reply_url = format!("{}/comment/{}", protocol_and_hostname, r.comment.id);
2020-12-15 19:39:18 +00:00
build_item(
&r.creator.name,
&r.comment.published,
&reply_url,
&r.comment.content,
protocol_and_hostname,
2020-12-15 19:39:18 +00:00
)
})
.collect::<Result<Vec<Item>, LemmyError>>()?;
let mut mention_items: Vec<Item> = mentions
.iter()
.map(|m| {
2023-06-14 09:34:29 +00:00
let mention_url = format!("{}/comment/{}", protocol_and_hostname, m.comment.id);
2020-12-16 16:09:21 +00:00
build_item(
&m.creator.name,
&m.comment.published,
&mention_url,
&m.comment.content,
protocol_and_hostname,
2020-12-16 16:09:21 +00:00
)
})
.collect::<Result<Vec<Item>, LemmyError>>()?;
reply_items.append(&mut mention_items);
Ok(reply_items)
}
#[tracing::instrument(skip_all)]
fn build_item(
creator_name: &str,
published: &DateTime<Utc>,
url: &str,
content: &str,
protocol_and_hostname: &str,
) -> Result<Item, LemmyError> {
// TODO add images
let author_url = format!("{protocol_and_hostname}/u/{creator_name}");
let guid = Some(Guid {
permalink: true,
value: url.to_owned(),
});
let description = Some(markdown_to_html(content));
Ok(Item {
title: Some(format!("Reply from {creator_name}")),
author: Some(format!(
"/u/{creator_name} <a href=\"{author_url}\">(link)</a>"
)),
pub_date: Some(published.to_rfc2822()),
comments: Some(url.to_owned()),
link: Some(url.to_owned()),
guid,
description,
..Default::default()
})
}
#[tracing::instrument(skip_all)]
fn create_post_items(
posts: Vec<PostView>,
protocol_and_hostname: &str,
) -> Result<Vec<Item>, LemmyError> {
2019-11-16 02:17:42 +00:00
let mut items: Vec<Item> = Vec::new();
for p in posts {
let post_url = format!("{}/post/{}", protocol_and_hostname, p.post.id);
let community_url = format!(
"{}/c/{}",
protocol_and_hostname,
sanitize_html(&p.community.name)
);
let dublin_core_ext = Some(DublinCoreExtension {
creators: vec![p.creator.actor_id.to_string()],
..DublinCoreExtension::default()
});
let guid = Some(Guid {
permalink: true,
value: post_url.clone(),
});
let mut description = format!("submitted by <a href=\"{}\">{}</a> to <a href=\"{}\">{}</a><br>{} points | <a href=\"{}\">{} comments</a>",
2020-12-11 15:27:33 +00:00
p.creator.actor_id,
sanitize_html(&p.creator.name),
community_url,
sanitize_html(&p.community.name),
2020-12-11 15:27:33 +00:00
p.counts.score,
post_url,
2020-12-11 15:27:33 +00:00
p.counts.comments);
// If its a url post, add it to the description
// and see if we can parse it as a media enclosure.
let enclosure_opt = p.post.url.map(|url| {
let link_html = format!("<br><a href=\"{url}\">{url}</a>");
description.push_str(&link_html);
let mime_type = p
.post
.url_content_type
.unwrap_or_else(|| "application/octet-stream".to_string());
let mut enclosure_bld = EnclosureBuilder::default();
enclosure_bld.url(url.as_str().to_string());
enclosure_bld.mime_type(mime_type);
enclosure_bld.length("0".to_string());
enclosure_bld.build()
});
2020-12-11 15:27:33 +00:00
if let Some(body) = p.post.body {
let html = markdown_to_html(&body);
description.push_str(&html);
2019-11-19 17:07:10 +00:00
}
let mut extensions = ExtensionMap::new();
// If there's a thumbnail URL, add a media:content tag to display it.
// See https://www.rssboard.org/media-rss#media-content for details.
if let Some(url) = p.post.thumbnail_url {
let mut thumbnail_ext = ExtensionBuilder::default();
thumbnail_ext.name("media:content".to_string());
thumbnail_ext.attrs(BTreeMap::from([
("url".to_string(), url.to_string()),
("medium".to_string(), "image".to_string()),
]));
extensions.insert(
"media".to_string(),
BTreeMap::from([("content".to_string(), vec![thumbnail_ext.build()])]),
);
}
let i = Item {
title: Some(sanitize_html(sanitize_xml(p.post.name).as_str())),
pub_date: Some(p.post.published.to_rfc2822()),
comments: Some(post_url.clone()),
guid,
description: Some(sanitize_xml(description)),
dublin_core_ext,
link: Some(post_url.clone()),
extensions,
enclosure: enclosure_opt,
..Default::default()
};
items.push(i);
2019-11-16 02:17:42 +00:00
}
Ok(items)
}