452 lines
15 KiB
Rust
452 lines
15 KiB
Rust
#![cfg_attr(feature = "cargo-clippy", allow(type_complexity))]
|
|
|
|
use gdk::FrameClockExt;
|
|
use gdk_pixbuf::{Object, Pixbuf};
|
|
use glib::{self, object::WeakRef};
|
|
use gtk;
|
|
use gtk::prelude::*;
|
|
use gtk::{IsA, Widget};
|
|
|
|
use chrono::prelude::*;
|
|
use crossbeam_channel::{bounded, unbounded, Sender};
|
|
use failure::Error;
|
|
use fragile::Fragile;
|
|
use rayon;
|
|
use regex::Regex;
|
|
use reqwest;
|
|
use serde_json::Value;
|
|
|
|
// use podcasts_data::feed;
|
|
use podcasts_data::dbqueries;
|
|
use podcasts_data::opml;
|
|
use podcasts_data::pipeline;
|
|
use podcasts_data::utils::checkup;
|
|
use podcasts_data::Source;
|
|
use podcasts_downloader::downloader;
|
|
|
|
use std::collections::{HashMap, HashSet};
|
|
use std::sync::{Arc, Mutex, RwLock};
|
|
|
|
use app::Action;
|
|
|
|
use i18n::i18n;
|
|
|
|
/// Lazy evaluates and loads widgets to the parent `container` widget.
|
|
///
|
|
/// Accepts an `IntoIterator`, `data`, as the source from which each widget
|
|
/// will be constructed. An `FnMut` function that returns the desired
|
|
/// widget should be passed as the widget `constructor`. You can also specify
|
|
/// a `callback` that will be executed when the iteration finish.
|
|
///
|
|
/// ```no_run
|
|
/// # struct Message;
|
|
/// # struct MessageWidget(gtk::Label);
|
|
///
|
|
/// # impl MessageWidget {
|
|
/// # fn new(_: Message) -> Self {
|
|
/// # MessageWidget(gtk::Label::new("A message"))
|
|
/// # }
|
|
/// # }
|
|
///
|
|
/// let messages: Vec<Message> = Vec::new();
|
|
/// let list = gtk::ListBox::new();
|
|
/// let constructor = |m| MessageWidget::new(m).0;
|
|
/// lazy_load(messages, list, constructor, || {});
|
|
/// ```
|
|
///
|
|
/// If you have already constructed the widgets and only want to
|
|
/// load them to the parent you can pass a closure that returns it's
|
|
/// own argument to the constructor.
|
|
///
|
|
/// ```no_run
|
|
/// # use std::collections::binary_heap::BinaryHeap;
|
|
/// let widgets: BinaryHeap<gtk::Button> = BinaryHeap::new();
|
|
/// let list = gtk::ListBox::new();
|
|
/// lazy_load(widgets, list, |w| w, || {});
|
|
/// ```
|
|
pub(crate) fn lazy_load<T, C, F, W, U>(
|
|
data: T,
|
|
container: WeakRef<C>,
|
|
mut contructor: F,
|
|
callback: U,
|
|
) where
|
|
T: IntoIterator + 'static,
|
|
T::Item: 'static,
|
|
C: IsA<Object> + ContainerExt + 'static,
|
|
F: FnMut(T::Item) -> W + 'static,
|
|
W: IsA<Widget> + WidgetExt,
|
|
U: Fn() + 'static,
|
|
{
|
|
let func = move |x| {
|
|
let container = match container.upgrade() {
|
|
Some(c) => c,
|
|
None => return,
|
|
};
|
|
|
|
let widget = contructor(x);
|
|
container.add(&widget);
|
|
widget.show();
|
|
};
|
|
lazy_load_full(data, func, callback);
|
|
}
|
|
|
|
/// Iterate over `data` and execute `func` using a `gtk::idle_add()`,
|
|
/// when the iteration finishes, it executes `finish_callback`.
|
|
///
|
|
/// This is a more flexible version of `lazy_load` with less constrains.
|
|
/// If you just want to lazy add `widgets` to a `container` check if
|
|
/// `lazy_load` fits your needs first.
|
|
#[cfg_attr(feature = "cargo-clippy", allow(redundant_closure))]
|
|
pub(crate) fn lazy_load_full<T, F, U>(data: T, mut func: F, finish_callback: U)
|
|
where
|
|
T: IntoIterator + 'static,
|
|
T::Item: 'static,
|
|
F: FnMut(T::Item) + 'static,
|
|
U: Fn() + 'static,
|
|
{
|
|
let mut data = data.into_iter();
|
|
gtk::idle_add(move || {
|
|
data.next()
|
|
.map(|x| func(x))
|
|
.map(|_| glib::Continue(true))
|
|
.unwrap_or_else(|| {
|
|
finish_callback();
|
|
glib::Continue(false)
|
|
})
|
|
});
|
|
}
|
|
|
|
// Kudos to Julian Sparber
|
|
// https://blogs.gnome.org/jsparber/2018/04/29/animate-a-scrolledwindow/
|
|
#[cfg_attr(feature = "cargo-clippy", allow(float_cmp))]
|
|
pub(crate) fn smooth_scroll_to(view: >k::ScrolledWindow, target: >k::Adjustment) {
|
|
if let Some(adj) = view.get_vadjustment() {
|
|
if let Some(clock) = view.get_frame_clock() {
|
|
let duration = 200;
|
|
let start = adj.get_value();
|
|
let end = target.get_value();
|
|
let start_time = clock.get_frame_time();
|
|
let end_time = start_time + 1000 * duration;
|
|
|
|
view.add_tick_callback(move |_, clock| {
|
|
let now = clock.get_frame_time();
|
|
// FIXME: `adj.get_value != end` is a float comparison...
|
|
if now < end_time && adj.get_value().abs() != end.abs() {
|
|
let mut t = (now - start_time) as f64 / (end_time - start_time) as f64;
|
|
t = ease_out_cubic(t);
|
|
adj.set_value(start + t * (end - start));
|
|
glib::Continue(true)
|
|
} else {
|
|
adj.set_value(end);
|
|
glib::Continue(false)
|
|
}
|
|
});
|
|
}
|
|
}
|
|
}
|
|
|
|
// From clutter-easing.c, based on Robert Penner's
|
|
// infamous easing equations, MIT license.
|
|
fn ease_out_cubic(t: f64) -> f64 {
|
|
let p = t - 1f64;
|
|
p * p * p + 1f64
|
|
}
|
|
|
|
lazy_static! {
|
|
static ref IGNORESHOWS: Arc<Mutex<HashSet<i32>>> = Arc::new(Mutex::new(HashSet::new()));
|
|
}
|
|
|
|
pub(crate) fn ignore_show(id: i32) -> Result<bool, Error> {
|
|
IGNORESHOWS
|
|
.lock()
|
|
.map(|mut guard| guard.insert(id))
|
|
.map_err(|err| format_err!("{}", err))
|
|
}
|
|
|
|
pub(crate) fn uningore_show(id: i32) -> Result<bool, Error> {
|
|
IGNORESHOWS
|
|
.lock()
|
|
.map(|mut guard| guard.remove(&id))
|
|
.map_err(|err| format_err!("{}", err))
|
|
}
|
|
|
|
pub(crate) fn get_ignored_shows() -> Result<Vec<i32>, Error> {
|
|
IGNORESHOWS
|
|
.lock()
|
|
.map(|guard| guard.iter().cloned().collect::<Vec<_>>())
|
|
.map_err(|err| format_err!("{}", err))
|
|
}
|
|
|
|
pub(crate) fn cleanup(cleanup_date: DateTime<Utc>) {
|
|
checkup(cleanup_date)
|
|
.map_err(|err| error!("Check up failed: {}", err))
|
|
.ok();
|
|
}
|
|
|
|
pub(crate) fn refresh<S>(source: Option<S>, sender: Sender<Action>)
|
|
where
|
|
S: IntoIterator<Item = Source> + Send + 'static,
|
|
{
|
|
// If we try to update the whole db,
|
|
// Exit early if `source` table is empty
|
|
if source.is_none() {
|
|
match dbqueries::is_source_populated(&[]) {
|
|
Ok(false) => {
|
|
info!("No source of feeds where found, returning");
|
|
return;
|
|
}
|
|
Err(err) => debug_assert!(false, err),
|
|
_ => (),
|
|
};
|
|
}
|
|
|
|
refresh_feed(source, sender)
|
|
}
|
|
|
|
/// Update the rss feed(s) originating from `source`.
|
|
/// If `source` is None, Fetches all the `Source` entries in the database and updates them.
|
|
fn refresh_feed<S>(source: Option<S>, sender: Sender<Action>)
|
|
where
|
|
S: IntoIterator<Item = Source> + Send + 'static,
|
|
{
|
|
rayon::spawn(move || {
|
|
let (up_sender, up_receiver) = bounded(1);
|
|
sender.send(Action::ShowUpdateNotif(up_receiver));
|
|
|
|
if let Some(s) = source {
|
|
// Refresh only specified feeds
|
|
pipeline::run(s)
|
|
.map_err(|err| error!("Error: {}", err))
|
|
.map_err(|_| error!("Error while trying to update the database."))
|
|
.ok();
|
|
} else {
|
|
// Refresh all the feeds
|
|
dbqueries::get_sources()
|
|
.map(|s| s.into_iter())
|
|
.and_then(pipeline::run)
|
|
.map_err(|err| error!("Error: {}", err))
|
|
.ok();
|
|
};
|
|
|
|
up_sender.send(true);
|
|
});
|
|
}
|
|
|
|
lazy_static! {
|
|
static ref CACHED_PIXBUFS: RwLock<HashMap<(i32, u32), Mutex<Fragile<Pixbuf>>>> =
|
|
{ RwLock::new(HashMap::new()) };
|
|
static ref COVER_DL_REGISTRY: RwLock<HashSet<i32>> = RwLock::new(HashSet::new());
|
|
static ref THREADPOOL: rayon::ThreadPool = rayon::ThreadPoolBuilder::new().build().unwrap();
|
|
}
|
|
|
|
// Since gdk_pixbuf::Pixbuf is reference counted and every episode,
|
|
// use the cover of the Podcast Feed/Show, We can only create a Pixbuf
|
|
// cover per show and pass around the Rc pointer.
|
|
//
|
|
// GObjects do not implement Send trait, so SendCell is a way around that.
|
|
// Also lazy_static requires Sync trait, so that's what the mutexes are.
|
|
// TODO: maybe use something that would just scale to requested size?
|
|
pub(crate) fn set_image_from_path(
|
|
image: >k::Image,
|
|
show_id: i32,
|
|
size: u32,
|
|
) -> Result<(), Error> {
|
|
// Check if there's an active download about this show cover.
|
|
// If there is, a callback will be set so this function will be called again.
|
|
// If the download succeeds, there should be a quick return from the pixbuf cache_image
|
|
// If it fails another download will be scheduled.
|
|
if let Ok(guard) = COVER_DL_REGISTRY.read() {
|
|
if guard.contains(&show_id) {
|
|
let callback = clone!(image => move || {
|
|
let _ = set_image_from_path(&image, show_id, size);
|
|
glib::Continue(false)
|
|
});
|
|
gtk::timeout_add(250, callback);
|
|
return Ok(());
|
|
}
|
|
}
|
|
|
|
if let Ok(hashmap) = CACHED_PIXBUFS.read() {
|
|
// Check if the requested (cover + size) is already in the cache
|
|
// and if so do an early return after that.
|
|
if let Some(guard) = hashmap.get(&(show_id, size)) {
|
|
guard
|
|
.lock()
|
|
.map_err(|err| format_err!("Fragile Mutex: {}", err))
|
|
.and_then(|fragile| {
|
|
fragile
|
|
.try_get()
|
|
.map(|px| image.set_from_pixbuf(px))
|
|
.map_err(From::from)
|
|
})?;
|
|
|
|
return Ok(());
|
|
}
|
|
}
|
|
|
|
let (sender, receiver) = unbounded();
|
|
THREADPOOL.spawn(move || {
|
|
if let Ok(mut guard) = COVER_DL_REGISTRY.write() {
|
|
guard.insert(show_id);
|
|
if let Ok(pd) = dbqueries::get_podcast_cover_from_id(show_id) {
|
|
sender.send(downloader::cache_image(&pd));
|
|
}
|
|
guard.remove(&show_id);
|
|
}
|
|
});
|
|
|
|
let image = image.clone();
|
|
let s = size as i32;
|
|
gtk::timeout_add(25, move || {
|
|
if let Some(path) = receiver.try_recv() {
|
|
if let Ok(path) = path {
|
|
if let Ok(px) = Pixbuf::new_from_file_at_scale(&path, s, s, true) {
|
|
if let Ok(mut hashmap) = CACHED_PIXBUFS.write() {
|
|
hashmap.insert((show_id, size), Mutex::new(Fragile::new(px.clone())));
|
|
image.set_from_pixbuf(&px);
|
|
}
|
|
}
|
|
}
|
|
glib::Continue(false)
|
|
} else {
|
|
glib::Continue(true)
|
|
}
|
|
});
|
|
Ok(())
|
|
}
|
|
|
|
// FIXME: the signature should be `fn foo(s: Url) -> Result<Url, Error>`
|
|
pub(crate) fn itunes_to_rss(url: &str) -> Result<String, Error> {
|
|
let id = itunes_id_from_url(url).ok_or_else(|| format_err!("Failed to find an iTunes ID."))?;
|
|
lookup_id(id)
|
|
}
|
|
|
|
fn itunes_id_from_url(url: &str) -> Option<u32> {
|
|
lazy_static! {
|
|
static ref RE: Regex = Regex::new(r"/id([0-9]+)").unwrap();
|
|
}
|
|
|
|
// Get the itunes id from the url
|
|
let foo = RE.captures_iter(url).nth(0)?.get(1)?.as_str();
|
|
// Parse it to a u32, this *should* never fail
|
|
foo.parse::<u32>().ok()
|
|
}
|
|
|
|
fn lookup_id(id: u32) -> Result<String, Error> {
|
|
let url = format!("https://itunes.apple.com/lookup?id={}&entity=podcast", id);
|
|
let req: Value = reqwest::get(&url)?.json()?;
|
|
let rssurl = || -> Option<&str> { req.get("results")?.get(0)?.get("feedUrl")?.as_str() };
|
|
rssurl()
|
|
.map(From::from)
|
|
.ok_or_else(|| format_err!("Failed to get url from itunes response"))
|
|
}
|
|
|
|
pub(crate) fn on_import_clicked(window: >k::ApplicationWindow, sender: &Sender<Action>) {
|
|
use glib::translate::ToGlib;
|
|
use gtk::{FileChooserAction, FileChooserNative, FileFilter, ResponseType};
|
|
|
|
// Create the FileChooser Dialog
|
|
let dialog = FileChooserNative::new(
|
|
Some(i18n("Select the file from which to you want to import shows.").as_str()),
|
|
Some(window),
|
|
FileChooserAction::Open,
|
|
Some(i18n("_Import").as_str()),
|
|
None,
|
|
);
|
|
|
|
// Do not show hidden(.thing) files
|
|
dialog.set_show_hidden(false);
|
|
|
|
// Set a filter to show only xml files
|
|
let filter = FileFilter::new();
|
|
FileFilterExt::set_name(&filter, Some(i18n("OPML file").as_str()));
|
|
filter.add_mime_type("application/xml");
|
|
filter.add_mime_type("text/xml");
|
|
dialog.add_filter(&filter);
|
|
|
|
dialog.connect_response(clone!(sender => move |dialog, resp| {
|
|
debug!("Dialog Response {}", resp);
|
|
if resp == ResponseType::Accept.to_glib() {
|
|
if let Some(filename) = dialog.get_filename() {
|
|
debug!("File selected: {:?}", filename);
|
|
|
|
rayon::spawn(clone!(sender => move || {
|
|
// Parse the file and import the feeds
|
|
if let Ok(sources) = opml::import_from_file(filename) {
|
|
// Refresh the successfully parsed feeds to index them
|
|
refresh(Some(sources), sender)
|
|
} else {
|
|
let text = i18n("Failed to parse the imported file");
|
|
sender.send(Action::ErrorNotification(text));
|
|
}
|
|
}))
|
|
} else {
|
|
let text = i18n("Selected file could not be accessed.");
|
|
sender.send(Action::ErrorNotification(text));
|
|
}
|
|
}
|
|
|
|
dialog.destroy();
|
|
}));
|
|
|
|
dialog.run();
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
use failure::Error;
|
|
// use podcasts_data::Source;
|
|
// use podcasts_data::dbqueries;
|
|
|
|
// #[test]
|
|
// This test inserts an rss feed to your `XDG_DATA/podcasts/podcasts.db` so we make it explicit
|
|
// to run it.
|
|
// #[ignore]
|
|
// Disabled till https://gitlab.gnome.org/World/podcasts/issues/56
|
|
// fn test_set_image_from_path() {
|
|
// let url = "https://web.archive.org/web/20180120110727if_/https://rss.acast.com/thetipoff";
|
|
// Create and index a source
|
|
// let source = Source::from_url(url).unwrap();
|
|
// Copy it's id
|
|
// let sid = source.id();
|
|
// pipeline::run(vec![source], true).unwrap();
|
|
|
|
// Get the Podcast
|
|
// let img = gtk::Image::new();
|
|
// let pd = dbqueries::get_podcast_from_source_id(sid).unwrap().into();
|
|
// let pxbuf = set_image_from_path(&img, Arc::new(pd), 256);
|
|
// assert!(pxbuf.is_ok());
|
|
// }
|
|
|
|
#[test]
|
|
fn test_itunes_to_rss() -> Result<(), Error> {
|
|
let itunes_url = "https://itunes.apple.com/podcast/id1195206601";
|
|
let rss_url = String::from("http://feeds.feedburner.com/InterceptedWithJeremyScahill");
|
|
assert_eq!(rss_url, itunes_to_rss(itunes_url)?);
|
|
|
|
let itunes_url = "https://itunes.apple.com/podcast/id000000000000000";
|
|
assert!(itunes_to_rss(itunes_url).is_err());
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn test_itunes_id() -> Result<(), Error> {
|
|
let id = 1195206601;
|
|
let itunes_url = "https://itunes.apple.com/podcast/id1195206601";
|
|
assert_eq!(id, itunes_id_from_url(itunes_url).unwrap());
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn test_itunes_lookup_id() -> Result<(), Error> {
|
|
let id = 1195206601;
|
|
let rss_url = "http://feeds.feedburner.com/InterceptedWithJeremyScahill";
|
|
assert_eq!(rss_url, lookup_id(id)?);
|
|
|
|
let id = 000000000;
|
|
assert!(lookup_id(id).is_err());
|
|
Ok(())
|
|
}
|
|
}
|