Compare commits
3 commits
b3331b36a7
...
c3f82f765b
Author | SHA1 | Date | |
---|---|---|---|
c3f82f765b | |||
29ad2f99d4 | |||
0008e305c2 |
16 changed files with 266 additions and 194 deletions
40
README.md
40
README.md
|
@ -1,5 +1,7 @@
|
||||||
# RustyPipe
|
# RustyPipe
|
||||||
|
|
||||||
|
[![CI status](https://ci.thetadev.de/api/badges/ThetaDev/rustypipe/status.svg)](https://ci.thetadev.de/ThetaDev/rustypipe)
|
||||||
|
|
||||||
Client for the public YouTube / YouTube Music API (Innertube),
|
Client for the public YouTube / YouTube Music API (Innertube),
|
||||||
inspired by [NewPipe](https://github.com/TeamNewPipe/NewPipeExtractor).
|
inspired by [NewPipe](https://github.com/TeamNewPipe/NewPipeExtractor).
|
||||||
|
|
||||||
|
@ -7,25 +9,25 @@ inspired by [NewPipe](https://github.com/TeamNewPipe/NewPipeExtractor).
|
||||||
|
|
||||||
### YouTube
|
### YouTube
|
||||||
|
|
||||||
- [X] **Player** (video/audio streams, subtitles)
|
- **Player** (video/audio streams, subtitles)
|
||||||
- [X] **Playlist**
|
- **Playlist**
|
||||||
- [X] **VideoDetails** (metadata, comments, recommended videos)
|
- **VideoDetails** (metadata, comments, recommended videos)
|
||||||
- [X] **Channel** (videos, shorts, livestreams, playlists, info, search)
|
- **Channel** (videos, shorts, livestreams, playlists, info, search)
|
||||||
- [X] **ChannelRSS**
|
- **ChannelRSS**
|
||||||
- [X] **Search** (with filters)
|
- **Search** (with filters)
|
||||||
- [X] **Search suggestions**
|
- **Search suggestions**
|
||||||
- [X] **Trending**
|
- **Trending**
|
||||||
- [X] **URL resolver**
|
- **URL resolver**
|
||||||
|
|
||||||
### YouTube Music
|
### YouTube Music
|
||||||
|
|
||||||
- [X] **Playlist**
|
- **Playlist**
|
||||||
- [X] **Album**
|
- **Album**
|
||||||
- [X] **Artist**
|
- **Artist**
|
||||||
- [X] **Search**
|
- **Search**
|
||||||
- [X] **Search suggestions**
|
- **Search suggestions**
|
||||||
- [X] **Radio**
|
- **Radio**
|
||||||
- [X] **Track details** (lyrics, recommendations)
|
- **Track details** (lyrics, recommendations)
|
||||||
- [X] **Moods/Genres**
|
- **Moods/Genres**
|
||||||
- [X] **Charts**
|
- **Charts**
|
||||||
- [X] **New**
|
- **New** (albums, music videos)
|
||||||
|
|
|
@ -8,7 +8,7 @@ use crate::{
|
||||||
error::{Error, ExtractionError},
|
error::{Error, ExtractionError},
|
||||||
model::{AlbumItem, ArtistId, MusicArtist},
|
model::{AlbumItem, ArtistId, MusicArtist},
|
||||||
serializer::MapResult,
|
serializer::MapResult,
|
||||||
util::{self, TryRemove},
|
util,
|
||||||
};
|
};
|
||||||
|
|
||||||
use super::{
|
use super::{
|
||||||
|
@ -331,9 +331,12 @@ impl MapResponse<Vec<AlbumItem>> for response::MusicArtistAlbums {
|
||||||
) -> Result<MapResult<Vec<AlbumItem>>, ExtractionError> {
|
) -> Result<MapResult<Vec<AlbumItem>>, ExtractionError> {
|
||||||
// dbg!(&self);
|
// dbg!(&self);
|
||||||
|
|
||||||
let mut content = self.contents.single_column_browse_results_renderer.contents;
|
let grids = self
|
||||||
let grids = content
|
.contents
|
||||||
.try_swap_remove(0)
|
.single_column_browse_results_renderer
|
||||||
|
.contents
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no content")))?
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no content")))?
|
||||||
.tab_renderer
|
.tab_renderer
|
||||||
.content
|
.content
|
||||||
|
|
|
@ -125,14 +125,17 @@ impl MapResponse<MusicPlaylist> for response::MusicPlaylist {
|
||||||
) -> Result<MapResult<MusicPlaylist>, ExtractionError> {
|
) -> Result<MapResult<MusicPlaylist>, ExtractionError> {
|
||||||
// dbg!(&self);
|
// dbg!(&self);
|
||||||
|
|
||||||
let mut content = self.contents.single_column_browse_results_renderer.contents;
|
let music_contents = self
|
||||||
let mut music_contents = content
|
.contents
|
||||||
.try_swap_remove(0)
|
.single_column_browse_results_renderer
|
||||||
|
.contents
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no content")))?
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no content")))?
|
||||||
.tab_renderer
|
.tab_renderer
|
||||||
.content
|
.content
|
||||||
.section_list_renderer;
|
.section_list_renderer;
|
||||||
let mut shelf = music_contents
|
let shelf = music_contents
|
||||||
.contents
|
.contents
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.find_map(|section| match section {
|
.find_map(|section| match section {
|
||||||
|
@ -157,7 +160,8 @@ impl MapResponse<MusicPlaylist> for response::MusicPlaylist {
|
||||||
|
|
||||||
let ctoken = shelf
|
let ctoken = shelf
|
||||||
.continuations
|
.continuations
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.map(|cont| cont.next_continuation_data.continuation);
|
.map(|cont| cont.next_continuation_data.continuation);
|
||||||
|
|
||||||
let track_count = if ctoken.is_some() {
|
let track_count = if ctoken.is_some() {
|
||||||
|
@ -177,7 +181,8 @@ impl MapResponse<MusicPlaylist> for response::MusicPlaylist {
|
||||||
|
|
||||||
let related_ctoken = music_contents
|
let related_ctoken = music_contents
|
||||||
.continuations
|
.continuations
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.map(|c| c.next_continuation_data.continuation);
|
.map(|c| c.next_continuation_data.continuation);
|
||||||
|
|
||||||
let (from_ytm, channel, name, thumbnail, description) = match self.header {
|
let (from_ytm, channel, name, thumbnail, description) = match self.header {
|
||||||
|
@ -269,9 +274,12 @@ impl MapResponse<MusicAlbum> for response::MusicPlaylist {
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no header")))?
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no header")))?
|
||||||
.music_detail_header_renderer;
|
.music_detail_header_renderer;
|
||||||
|
|
||||||
let mut content = self.contents.single_column_browse_results_renderer.contents;
|
let sections = self
|
||||||
let sections = content
|
.contents
|
||||||
.try_swap_remove(0)
|
.single_column_browse_results_renderer
|
||||||
|
.contents
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no content")))?
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no content")))?
|
||||||
.tab_renderer
|
.tab_renderer
|
||||||
.content
|
.content
|
||||||
|
@ -320,7 +328,8 @@ impl MapResponse<MusicAlbum> for response::MusicPlaylist {
|
||||||
|
|
||||||
let (artists, by_va) = map_artists(artists_p);
|
let (artists, by_va) = map_artists(artists_p);
|
||||||
let album_type_txt = subtitle_split
|
let album_type_txt = subtitle_split
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.map(|part| part.to_string())
|
.map(|part| part.to_string())
|
||||||
.unwrap_or_default();
|
.unwrap_or_default();
|
||||||
|
|
||||||
|
@ -329,12 +338,13 @@ impl MapResponse<MusicAlbum> for response::MusicPlaylist {
|
||||||
|
|
||||||
let (artist_id, playlist_id) = header
|
let (artist_id, playlist_id) = header
|
||||||
.menu
|
.menu
|
||||||
.map(|mut menu| {
|
.map(|menu| {
|
||||||
(
|
(
|
||||||
map_artist_id(menu.menu_renderer.items),
|
map_artist_id(menu.menu_renderer.items),
|
||||||
menu.menu_renderer
|
menu.menu_renderer
|
||||||
.top_level_buttons
|
.top_level_buttons
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.map(|btn| {
|
.map(|btn| {
|
||||||
btn.button_renderer
|
btn.button_renderer
|
||||||
.navigation_endpoint
|
.navigation_endpoint
|
||||||
|
|
|
@ -10,7 +10,6 @@ use crate::{
|
||||||
MusicSearchFiltered, MusicSearchResult, MusicSearchSuggestion, TrackItem,
|
MusicSearchFiltered, MusicSearchResult, MusicSearchSuggestion, TrackItem,
|
||||||
},
|
},
|
||||||
serializer::MapResult,
|
serializer::MapResult,
|
||||||
util::TryRemove,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
use super::{response, ClientType, MapResponse, RustyPipeQuery, YTContext};
|
use super::{response, ClientType, MapResponse, RustyPipeQuery, YTContext};
|
||||||
|
@ -234,9 +233,12 @@ impl MapResponse<MusicSearchResult> for response::MusicSearch {
|
||||||
) -> Result<MapResult<MusicSearchResult>, crate::error::ExtractionError> {
|
) -> Result<MapResult<MusicSearchResult>, crate::error::ExtractionError> {
|
||||||
// dbg!(&self);
|
// dbg!(&self);
|
||||||
|
|
||||||
let mut tabs = self.contents.tabbed_search_results_renderer.contents;
|
let sections = self
|
||||||
let sections = tabs
|
.contents
|
||||||
.try_swap_remove(0)
|
.tabbed_search_results_renderer
|
||||||
|
.contents
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no tab")))?
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no tab")))?
|
||||||
.tab_renderer
|
.tab_renderer
|
||||||
.content
|
.content
|
||||||
|
@ -262,8 +264,8 @@ impl MapResponse<MusicSearchResult> for response::MusicSearch {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
response::music_search::ItemSection::ItemSectionRenderer { mut contents } => {
|
response::music_search::ItemSection::ItemSectionRenderer { contents } => {
|
||||||
if let Some(corrected) = contents.try_swap_remove(0) {
|
if let Some(corrected) = contents.into_iter().next() {
|
||||||
corrected_query = Some(corrected.showing_results_for_renderer.corrected_query)
|
corrected_query = Some(corrected.showing_results_for_renderer.corrected_query)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -295,9 +297,10 @@ impl<T: FromYtItem> MapResponse<MusicSearchFiltered<T>> for response::MusicSearc
|
||||||
) -> Result<MapResult<MusicSearchFiltered<T>>, ExtractionError> {
|
) -> Result<MapResult<MusicSearchFiltered<T>>, ExtractionError> {
|
||||||
// dbg!(&self);
|
// dbg!(&self);
|
||||||
|
|
||||||
let mut tabs = self.contents.tabbed_search_results_renderer.contents;
|
let tabs = self.contents.tabbed_search_results_renderer.contents;
|
||||||
let sections = tabs
|
let sections = tabs
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no tab")))?
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no tab")))?
|
||||||
.tab_renderer
|
.tab_renderer
|
||||||
.content
|
.content
|
||||||
|
@ -309,17 +312,17 @@ impl<T: FromYtItem> MapResponse<MusicSearchFiltered<T>> for response::MusicSearc
|
||||||
let mut mapper = MusicListMapper::new(lang);
|
let mut mapper = MusicListMapper::new(lang);
|
||||||
|
|
||||||
sections.into_iter().for_each(|section| match section {
|
sections.into_iter().for_each(|section| match section {
|
||||||
response::music_search::ItemSection::MusicShelfRenderer(mut shelf) => {
|
response::music_search::ItemSection::MusicShelfRenderer(shelf) => {
|
||||||
mapper.map_response(shelf.contents);
|
mapper.map_response(shelf.contents);
|
||||||
if let Some(cont) = shelf.continuations.try_swap_remove(0) {
|
if let Some(cont) = shelf.continuations.into_iter().next() {
|
||||||
ctoken = Some(cont.next_continuation_data.continuation);
|
ctoken = Some(cont.next_continuation_data.continuation);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
response::music_search::ItemSection::MusicCardShelfRenderer(card) => {
|
response::music_search::ItemSection::MusicCardShelfRenderer(card) => {
|
||||||
mapper.map_card(card);
|
mapper.map_card(card);
|
||||||
}
|
}
|
||||||
response::music_search::ItemSection::ItemSectionRenderer { mut contents } => {
|
response::music_search::ItemSection::ItemSectionRenderer { contents } => {
|
||||||
if let Some(corrected) = contents.try_swap_remove(0) {
|
if let Some(corrected) = contents.into_iter().next() {
|
||||||
corrected_query = Some(corrected.showing_results_for_renderer.corrected_query)
|
corrected_query = Some(corrected.showing_results_for_renderer.corrected_query)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -404,7 +407,7 @@ mod tests {
|
||||||
#[case::default("default")]
|
#[case::default("default")]
|
||||||
#[case::typo("typo")]
|
#[case::typo("typo")]
|
||||||
#[case::radio("radio")]
|
#[case::radio("radio")]
|
||||||
#[case::radio("artist")]
|
#[case::artist("artist")]
|
||||||
fn map_music_search_main(#[case] name: &str) {
|
fn map_music_search_main(#[case] name: &str) {
|
||||||
let json_path = path!(*TESTFILES / "music_search" / format!("main_{name}.json"));
|
let json_path = path!(*TESTFILES / "music_search" / format!("main_{name}.json"));
|
||||||
let json_file = File::open(json_path).unwrap();
|
let json_file = File::open(json_path).unwrap();
|
||||||
|
|
|
@ -5,7 +5,6 @@ use crate::model::{
|
||||||
Comment, MusicItem, PlaylistVideo, YouTubeItem,
|
Comment, MusicItem, PlaylistVideo, YouTubeItem,
|
||||||
};
|
};
|
||||||
use crate::serializer::MapResult;
|
use crate::serializer::MapResult;
|
||||||
use crate::util::TryRemove;
|
|
||||||
|
|
||||||
use super::response::music_item::{map_queue_item, MusicListMapper, PlaylistPanelVideo};
|
use super::response::music_item::{map_queue_item, MusicListMapper, PlaylistPanelVideo};
|
||||||
use super::{response, ClientType, MapResponse, QContinuation, RustyPipeQuery};
|
use super::{response, ClientType, MapResponse, QContinuation, RustyPipeQuery};
|
||||||
|
@ -100,9 +99,10 @@ impl MapResponse<Paginator<YouTubeItem>> for response::Continuation {
|
||||||
) -> Result<MapResult<Paginator<YouTubeItem>>, ExtractionError> {
|
) -> Result<MapResult<Paginator<YouTubeItem>>, ExtractionError> {
|
||||||
let items = self
|
let items = self
|
||||||
.on_response_received_actions
|
.on_response_received_actions
|
||||||
.and_then(|mut actions| {
|
.and_then(|actions| {
|
||||||
actions
|
actions
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.map(|action| action.append_continuation_items_action.continuation_items)
|
.map(|action| action.append_continuation_items_action.continuation_items)
|
||||||
})
|
})
|
||||||
.or_else(|| {
|
.or_else(|| {
|
||||||
|
@ -168,7 +168,8 @@ impl MapResponse<Paginator<MusicItem>> for response::MusicContinuation {
|
||||||
|
|
||||||
let map_res = mapper.items();
|
let map_res = mapper.items();
|
||||||
let ctoken = continuations
|
let ctoken = continuations
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.map(|cont| cont.next_continuation_data.continuation);
|
.map(|cont| cont.next_continuation_data.continuation);
|
||||||
|
|
||||||
Ok(MapResult {
|
Ok(MapResult {
|
||||||
|
|
|
@ -65,10 +65,11 @@ impl MapResponse<Playlist> for response::Playlist {
|
||||||
_ => return Err(response::alerts_to_err(self.alerts)),
|
_ => return Err(response::alerts_to_err(self.alerts)),
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut tcbr_contents = contents.two_column_browse_results_renderer.contents;
|
let video_items = contents
|
||||||
|
.two_column_browse_results_renderer
|
||||||
let video_items = tcbr_contents
|
.contents
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed(
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed(
|
||||||
"twoColumnBrowseResultsRenderer empty",
|
"twoColumnBrowseResultsRenderer empty",
|
||||||
)))?
|
)))?
|
||||||
|
@ -76,13 +77,15 @@ impl MapResponse<Playlist> for response::Playlist {
|
||||||
.content
|
.content
|
||||||
.section_list_renderer
|
.section_list_renderer
|
||||||
.contents
|
.contents
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed(
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed(
|
||||||
"sectionListRenderer empty",
|
"sectionListRenderer empty",
|
||||||
)))?
|
)))?
|
||||||
.item_section_renderer
|
.item_section_renderer
|
||||||
.contents
|
.contents
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed(
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed(
|
||||||
"itemSectionRenderer empty",
|
"itemSectionRenderer empty",
|
||||||
)))?
|
)))?
|
||||||
|
@ -93,10 +96,11 @@ impl MapResponse<Playlist> for response::Playlist {
|
||||||
|
|
||||||
let (thumbnails, last_update_txt) = match self.sidebar {
|
let (thumbnails, last_update_txt) = match self.sidebar {
|
||||||
Some(sidebar) => {
|
Some(sidebar) => {
|
||||||
let mut sidebar_items = sidebar.playlist_sidebar_renderer.contents;
|
let sidebar_items = sidebar.playlist_sidebar_renderer.contents;
|
||||||
let mut primary =
|
let mut primary =
|
||||||
sidebar_items
|
sidebar_items
|
||||||
.try_swap_remove(0)
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed(
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed(
|
||||||
"no primary sidebar",
|
"no primary sidebar",
|
||||||
)))?;
|
)))?;
|
||||||
|
|
|
@ -11,7 +11,7 @@ use crate::{
|
||||||
text::{Text, TextComponents},
|
text::{Text, TextComponents},
|
||||||
MapResult,
|
MapResult,
|
||||||
},
|
},
|
||||||
util::{self, dictionary, TryRemove},
|
util::{self, dictionary},
|
||||||
};
|
};
|
||||||
|
|
||||||
use super::{
|
use super::{
|
||||||
|
@ -587,14 +587,14 @@ impl MusicListMapper {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Playlist item
|
// Playlist item
|
||||||
FlexColumnDisplayStyle::Default => {
|
FlexColumnDisplayStyle::Default => (
|
||||||
let mut fixed_columns = item.fixed_columns;
|
c2.map(TextComponents::from),
|
||||||
(
|
c3.map(TextComponents::from),
|
||||||
c2.map(TextComponents::from),
|
item.fixed_columns
|
||||||
c3.map(TextComponents::from),
|
.into_iter()
|
||||||
fixed_columns.try_swap_remove(0).map(TextComponents::from),
|
.next()
|
||||||
)
|
.map(TextComponents::from),
|
||||||
}
|
),
|
||||||
};
|
};
|
||||||
|
|
||||||
let duration =
|
let duration =
|
||||||
|
|
|
@ -477,7 +477,7 @@ impl<T> YouTubeListMapper<T> {
|
||||||
is_upcoming: video.upcoming_event_data.is_some(),
|
is_upcoming: video.upcoming_event_data.is_some(),
|
||||||
short_description: video
|
short_description: video
|
||||||
.detailed_metadata_snippets
|
.detailed_metadata_snippets
|
||||||
.and_then(|mut snippets| snippets.try_swap_remove(0).map(|s| s.snippet_text))
|
.and_then(|snippets| snippets.into_iter().next().map(|s| s.snippet_text))
|
||||||
.or(video.description_snippet),
|
.or(video.description_snippet),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -5,7 +5,6 @@ use crate::{
|
||||||
model::{paginator::Paginator, VideoItem},
|
model::{paginator::Paginator, VideoItem},
|
||||||
param::Language,
|
param::Language,
|
||||||
serializer::MapResult,
|
serializer::MapResult,
|
||||||
util::TryRemove,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
use super::{response, ClientType, MapResponse, QBrowse, QBrowseParams, RustyPipeQuery};
|
use super::{response, ClientType, MapResponse, QBrowse, QBrowseParams, RustyPipeQuery};
|
||||||
|
@ -56,9 +55,12 @@ impl MapResponse<Paginator<VideoItem>> for response::Startpage {
|
||||||
lang: crate::param::Language,
|
lang: crate::param::Language,
|
||||||
_deobf: Option<&crate::deobfuscate::DeobfData>,
|
_deobf: Option<&crate::deobfuscate::DeobfData>,
|
||||||
) -> Result<MapResult<Paginator<VideoItem>>, ExtractionError> {
|
) -> Result<MapResult<Paginator<VideoItem>>, ExtractionError> {
|
||||||
let mut contents = self.contents.two_column_browse_results_renderer.contents;
|
let grid = self
|
||||||
let grid = contents
|
.contents
|
||||||
.try_swap_remove(0)
|
.two_column_browse_results_renderer
|
||||||
|
.contents
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no contents")))?
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no contents")))?
|
||||||
.tab_renderer
|
.tab_renderer
|
||||||
.content
|
.content
|
||||||
|
@ -80,9 +82,12 @@ impl MapResponse<Vec<VideoItem>> for response::Trending {
|
||||||
lang: crate::param::Language,
|
lang: crate::param::Language,
|
||||||
_deobf: Option<&crate::deobfuscate::DeobfData>,
|
_deobf: Option<&crate::deobfuscate::DeobfData>,
|
||||||
) -> Result<MapResult<Vec<VideoItem>>, ExtractionError> {
|
) -> Result<MapResult<Vec<VideoItem>>, ExtractionError> {
|
||||||
let mut contents = self.contents.two_column_browse_results_renderer.contents;
|
let items = self
|
||||||
let items = contents
|
.contents
|
||||||
.try_swap_remove(0)
|
.two_column_browse_results_renderer
|
||||||
|
.contents
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no contents")))?
|
.ok_or(ExtractionError::InvalidData(Cow::Borrowed("no contents")))?
|
||||||
.tab_renderer
|
.tab_renderer
|
||||||
.content
|
.content
|
||||||
|
|
|
@ -129,11 +129,11 @@ impl MapResponse<VideoDetails> for response::VideoDetails {
|
||||||
}
|
}
|
||||||
response::video_details::VideoResultsItem::ItemSectionRenderer(section) => {
|
response::video_details::VideoResultsItem::ItemSectionRenderer(section) => {
|
||||||
match section {
|
match section {
|
||||||
response::video_details::ItemSection::CommentsEntryPoint { mut contents } => {
|
response::video_details::ItemSection::CommentsEntryPoint { contents } => {
|
||||||
comment_count_section = contents.try_swap_remove(0);
|
comment_count_section = contents.into_iter().next();
|
||||||
}
|
}
|
||||||
response::video_details::ItemSection::CommentItemSection { mut contents } => {
|
response::video_details::ItemSection::CommentItemSection { contents } => {
|
||||||
comment_ctoken_section = contents.try_swap_remove(0);
|
comment_ctoken_section = contents.into_iter().next();
|
||||||
}
|
}
|
||||||
response::video_details::ItemSection::None => {}
|
response::video_details::ItemSection::None => {}
|
||||||
}
|
}
|
||||||
|
|
|
@ -393,13 +393,14 @@ pub(crate) fn entry(lang: Language) -> Entry {
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
number_nd_tokens: ::phf::Map {
|
number_nd_tokens: ::phf::Map {
|
||||||
key: 15467950696543387533,
|
key: 12913932095322966823,
|
||||||
disps: &[
|
disps: &[
|
||||||
(0, 0),
|
(0, 0),
|
||||||
],
|
],
|
||||||
entries: &[
|
entries: &[
|
||||||
("ন\u{9be}ই", 0),
|
|
||||||
("১", 1),
|
("১", 1),
|
||||||
|
("ন\u{9be}ই", 0),
|
||||||
|
("১ট\u{9be}", 1),
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
album_types: ::phf::Map {
|
album_types: ::phf::Map {
|
||||||
|
@ -4659,6 +4660,7 @@ pub(crate) fn entry(lang: Language) -> Entry {
|
||||||
],
|
],
|
||||||
entries: &[
|
entries: &[
|
||||||
("ingen", 0),
|
("ingen", 0),
|
||||||
|
("én", 1),
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
album_types: ::phf::Map {
|
album_types: ::phf::Map {
|
||||||
|
@ -5032,8 +5034,10 @@ pub(crate) fn entry(lang: Language) -> Entry {
|
||||||
number_nd_tokens: ::phf::Map {
|
number_nd_tokens: ::phf::Map {
|
||||||
key: 12913932095322966823,
|
key: 12913932095322966823,
|
||||||
disps: &[
|
disps: &[
|
||||||
|
(0, 0),
|
||||||
],
|
],
|
||||||
entries: &[
|
entries: &[
|
||||||
|
("um", 1),
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
album_types: ::phf::Map {
|
album_types: ::phf::Map {
|
||||||
|
|
112
src/util/mod.rs
112
src/util/mod.rs
|
@ -10,7 +10,7 @@ pub use protobuf::{string_from_pb, ProtoBuilder};
|
||||||
use std::{
|
use std::{
|
||||||
borrow::{Borrow, Cow},
|
borrow::{Borrow, Cow},
|
||||||
collections::BTreeMap,
|
collections::BTreeMap,
|
||||||
str::FromStr,
|
str::{FromStr, SplitWhitespace},
|
||||||
};
|
};
|
||||||
|
|
||||||
use base64::Engine;
|
use base64::Engine;
|
||||||
|
@ -331,36 +331,18 @@ where
|
||||||
}
|
}
|
||||||
|
|
||||||
if digits.is_empty() {
|
if digits.is_empty() {
|
||||||
if by_char {
|
SplitTokens::new(&filtered, by_char)
|
||||||
filtered
|
.find_map(|token| dict_entry.number_nd_tokens.get(token))
|
||||||
.chars()
|
.and_then(|n| (*n as u64).try_into().ok())
|
||||||
.find_map(|c| dict_entry.number_nd_tokens.get(&c.to_string()))
|
|
||||||
.and_then(|n| (*n as u64).try_into().ok())
|
|
||||||
} else {
|
|
||||||
filtered
|
|
||||||
.split_whitespace()
|
|
||||||
.find_map(|token| dict_entry.number_nd_tokens.get(token))
|
|
||||||
.and_then(|n| (*n as u64).try_into().ok())
|
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
let num = digits.parse::<u64>().ok()?;
|
let num = digits.parse::<u64>().ok()?;
|
||||||
|
|
||||||
let lookup_token = |token: &str| match token {
|
exp += SplitTokens::new(&filtered, by_char)
|
||||||
"k" => Some(3),
|
.filter_map(|token| match token {
|
||||||
_ => dict_entry.number_tokens.get(token).map(|t| *t as i32),
|
"k" => Some(3),
|
||||||
};
|
_ => dict_entry.number_tokens.get(token).map(|t| *t as i32),
|
||||||
|
})
|
||||||
if by_char {
|
.sum::<i32>();
|
||||||
exp += filtered
|
|
||||||
.chars()
|
|
||||||
.filter_map(|token| lookup_token(&token.to_string()))
|
|
||||||
.sum::<i32>();
|
|
||||||
} else {
|
|
||||||
exp += filtered
|
|
||||||
.split_whitespace()
|
|
||||||
.filter_map(lookup_token)
|
|
||||||
.sum::<i32>();
|
|
||||||
}
|
|
||||||
|
|
||||||
F::try_from(num.checked_mul((10_u64).checked_pow(exp.try_into().ok()?)?)?).ok()
|
F::try_from(num.checked_mul((10_u64).checked_pow(exp.try_into().ok()?)?)?).ok()
|
||||||
}
|
}
|
||||||
|
@ -415,6 +397,62 @@ pub fn b64_decode<T: AsRef<[u8]>>(input: T) -> Result<Vec<u8>, base64::DecodeErr
|
||||||
base64::engine::general_purpose::STANDARD.decode(input)
|
base64::engine::general_purpose::STANDARD.decode(input)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// An iterator over the chars in a string (in str format)
|
||||||
|
pub struct SplitChar<'a> {
|
||||||
|
txt: &'a str,
|
||||||
|
index: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> From<&'a str> for SplitChar<'a> {
|
||||||
|
fn from(value: &'a str) -> Self {
|
||||||
|
Self {
|
||||||
|
txt: value,
|
||||||
|
index: 0,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> Iterator for SplitChar<'a> {
|
||||||
|
type Item = &'a str;
|
||||||
|
|
||||||
|
fn next(&mut self) -> Option<Self::Item> {
|
||||||
|
self.txt
|
||||||
|
.get(self.index..)
|
||||||
|
.and_then(|txt| txt.chars().next())
|
||||||
|
.map(|c| {
|
||||||
|
let start = self.index;
|
||||||
|
self.index += c.len_utf8();
|
||||||
|
&self.txt[start..self.index]
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// An iterator for parsing strings. It can either iterate over words or characters.
|
||||||
|
pub enum SplitTokens<'a> {
|
||||||
|
Word(SplitWhitespace<'a>),
|
||||||
|
Char(SplitChar<'a>),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> SplitTokens<'a> {
|
||||||
|
pub fn new(s: &'a str, by_char: bool) -> Self {
|
||||||
|
match by_char {
|
||||||
|
true => Self::Char(SplitChar::from(s)),
|
||||||
|
false => Self::Word(s.split_whitespace()),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> Iterator for SplitTokens<'a> {
|
||||||
|
type Item = &'a str;
|
||||||
|
|
||||||
|
fn next(&mut self) -> Option<Self::Item> {
|
||||||
|
match self {
|
||||||
|
SplitTokens::Word(iter) => iter.next(),
|
||||||
|
SplitTokens::Char(iter) => iter.next(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
pub(crate) mod tests {
|
pub(crate) mod tests {
|
||||||
use std::{fs::File, io::BufReader, path::PathBuf};
|
use std::{fs::File, io::BufReader, path::PathBuf};
|
||||||
|
@ -550,4 +588,22 @@ pub(crate) mod tests {
|
||||||
let res = parse_large_numstr::<u64>(string, lang).expect(&emsg);
|
let res = parse_large_numstr::<u64>(string, lang).expect(&emsg);
|
||||||
assert_eq!(res, rounded, "{emsg}");
|
assert_eq!(res, rounded, "{emsg}");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn split_char() {
|
||||||
|
let teststr = "abc今天更新def";
|
||||||
|
let res = SplitTokens::new(teststr, true).collect::<Vec<_>>();
|
||||||
|
assert_eq!(res.len(), 10);
|
||||||
|
let res_str = res.into_iter().collect::<String>();
|
||||||
|
assert_eq!(res_str, teststr)
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn split_words() {
|
||||||
|
let teststr = "abc 今天更新 ghi";
|
||||||
|
let res = SplitTokens::new(teststr, false).collect::<Vec<_>>();
|
||||||
|
assert_eq!(res.len(), 3);
|
||||||
|
let res_str = res.join(" ");
|
||||||
|
assert_eq!(res_str, teststr)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -17,7 +17,7 @@ use time::{Date, Duration, Month, OffsetDateTime};
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
param::Language,
|
param::Language,
|
||||||
util::{self, dictionary},
|
util::{self, dictionary, SplitTokens},
|
||||||
};
|
};
|
||||||
|
|
||||||
/// Parsed TimeAgo string, contains amount and time unit.
|
/// Parsed TimeAgo string, contains amount and time unit.
|
||||||
|
@ -149,79 +149,39 @@ fn filter_str(string: &str) -> String {
|
||||||
.collect()
|
.collect()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn parse_ta_token(
|
struct TaTokenParser<'a> {
|
||||||
entry: &dictionary::Entry,
|
iter: SplitTokens<'a>,
|
||||||
by_char: bool,
|
tokens: &'a phf::Map<&'static str, TaToken>,
|
||||||
nd: bool,
|
}
|
||||||
filtered_str: &str,
|
|
||||||
) -> Option<TimeAgo> {
|
|
||||||
let tokens = match nd {
|
|
||||||
true => &entry.timeago_nd_tokens,
|
|
||||||
false => &entry.timeago_tokens,
|
|
||||||
};
|
|
||||||
let mut qu = 1;
|
|
||||||
|
|
||||||
if by_char {
|
impl<'a> TaTokenParser<'a> {
|
||||||
filtered_str.chars().find_map(|word| {
|
fn new(entry: &'a dictionary::Entry, by_char: bool, nd: bool, filtered_str: &'a str) -> Self {
|
||||||
tokens.get(&word.to_string()).and_then(|t| match t.unit {
|
let tokens = match nd {
|
||||||
Some(unit) => Some(TimeAgo { n: t.n * qu, unit }),
|
true => &entry.timeago_nd_tokens,
|
||||||
None => {
|
false => &entry.timeago_tokens,
|
||||||
qu = t.n;
|
};
|
||||||
None
|
Self {
|
||||||
}
|
iter: SplitTokens::new(filtered_str, by_char),
|
||||||
})
|
tokens,
|
||||||
})
|
}
|
||||||
} else {
|
|
||||||
filtered_str.split_whitespace().find_map(|word| {
|
|
||||||
tokens.get(word).and_then(|t| match t.unit {
|
|
||||||
Some(unit) => Some(TimeAgo { n: t.n * qu, unit }),
|
|
||||||
None => {
|
|
||||||
qu = t.n;
|
|
||||||
None
|
|
||||||
}
|
|
||||||
})
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn parse_ta_tokens(
|
impl<'a> Iterator for TaTokenParser<'a> {
|
||||||
entry: &dictionary::Entry,
|
type Item = TimeAgo;
|
||||||
by_char: bool,
|
|
||||||
nd: bool,
|
|
||||||
filtered_str: &str,
|
|
||||||
) -> Vec<TimeAgo> {
|
|
||||||
let tokens = match nd {
|
|
||||||
true => &entry.timeago_nd_tokens,
|
|
||||||
false => &entry.timeago_tokens,
|
|
||||||
};
|
|
||||||
let mut qu = 1;
|
|
||||||
|
|
||||||
if by_char {
|
fn next(&mut self) -> Option<Self::Item> {
|
||||||
filtered_str
|
// Quantity for parsing separate quantity + unit tokens
|
||||||
.chars()
|
let mut qu = 1;
|
||||||
.filter_map(|word| {
|
self.iter.find_map(|word| {
|
||||||
tokens.get(&word.to_string()).and_then(|t| match t.unit {
|
self.tokens.get(word).and_then(|t| match t.unit {
|
||||||
Some(unit) => Some(TimeAgo { n: t.n * qu, unit }),
|
Some(unit) => Some(TimeAgo { n: t.n * qu, unit }),
|
||||||
None => {
|
None => {
|
||||||
qu = t.n;
|
qu = t.n;
|
||||||
None
|
None
|
||||||
}
|
}
|
||||||
})
|
|
||||||
})
|
})
|
||||||
.collect()
|
})
|
||||||
} else {
|
|
||||||
filtered_str
|
|
||||||
.split_whitespace()
|
|
||||||
.filter_map(|word| {
|
|
||||||
tokens.get(word).and_then(|t| match t.unit {
|
|
||||||
Some(unit) => Some(TimeAgo { n: t.n * qu, unit }),
|
|
||||||
None => {
|
|
||||||
qu = t.n;
|
|
||||||
None
|
|
||||||
}
|
|
||||||
})
|
|
||||||
})
|
|
||||||
.collect()
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -240,7 +200,9 @@ pub fn parse_timeago(lang: Language, textual_date: &str) -> Option<TimeAgo> {
|
||||||
|
|
||||||
let qu: u8 = util::parse_numeric(textual_date).unwrap_or(1);
|
let qu: u8 = util::parse_numeric(textual_date).unwrap_or(1);
|
||||||
|
|
||||||
parse_ta_token(&entry, util::lang_by_char(lang), false, &filtered_str).map(|ta| ta * qu)
|
TaTokenParser::new(&entry, util::lang_by_char(lang), false, &filtered_str)
|
||||||
|
.next()
|
||||||
|
.map(|ta| ta * qu)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Parse a TimeAgo string (e.g. "29 minutes ago") into a Chrono DateTime object.
|
/// Parse a TimeAgo string (e.g. "29 minutes ago") into a Chrono DateTime object.
|
||||||
|
@ -273,11 +235,14 @@ pub fn parse_textual_date(lang: Language, textual_date: &str) -> Option<ParsedDa
|
||||||
let nums = util::parse_numeric_vec::<u16>(textual_date);
|
let nums = util::parse_numeric_vec::<u16>(textual_date);
|
||||||
|
|
||||||
match nums.len() {
|
match nums.len() {
|
||||||
0 => match parse_ta_token(&entry, by_char, true, &filtered_str) {
|
0 => match TaTokenParser::new(&entry, by_char, true, &filtered_str).next() {
|
||||||
Some(timeago) => Some(ParsedDate::Relative(timeago)),
|
Some(timeago) => Some(ParsedDate::Relative(timeago)),
|
||||||
None => parse_ta_token(&entry, by_char, false, &filtered_str).map(ParsedDate::Relative),
|
None => TaTokenParser::new(&entry, by_char, false, &filtered_str)
|
||||||
|
.next()
|
||||||
|
.map(ParsedDate::Relative),
|
||||||
},
|
},
|
||||||
1 => parse_ta_token(&entry, by_char, false, &filtered_str)
|
1 => TaTokenParser::new(&entry, by_char, false, &filtered_str)
|
||||||
|
.next()
|
||||||
.map(|timeago| ParsedDate::Relative(timeago * nums[0] as u8)),
|
.map(|timeago| ParsedDate::Relative(timeago * nums[0] as u8)),
|
||||||
2..=3 => {
|
2..=3 => {
|
||||||
if nums.len() == entry.date_order.len() {
|
if nums.len() == entry.date_order.len() {
|
||||||
|
@ -348,12 +313,10 @@ pub fn parse_video_duration(lang: Language, video_duration: &str) -> Option<u32>
|
||||||
} else {
|
} else {
|
||||||
part.digits.parse::<u32>().ok()?
|
part.digits.parse::<u32>().ok()?
|
||||||
};
|
};
|
||||||
let tokens = parse_ta_tokens(&entry, by_char, false, &part.word);
|
let mut tokens = TaTokenParser::new(&entry, by_char, false, &part.word).peekable();
|
||||||
if tokens.is_empty() {
|
tokens.peek()?;
|
||||||
return None;
|
|
||||||
}
|
|
||||||
|
|
||||||
tokens.iter().for_each(|ta| {
|
tokens.for_each(|ta| {
|
||||||
secs += n * ta.secs() as u32;
|
secs += n * ta.secs() as u32;
|
||||||
n = 1;
|
n = 1;
|
||||||
});
|
});
|
||||||
|
@ -805,4 +768,12 @@ mod tests {
|
||||||
let now = OffsetDateTime::now_utc();
|
let now = OffsetDateTime::now_utc();
|
||||||
assert_eq!(date.year(), now.year() - 1);
|
assert_eq!(date.year(), now.year() - 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn tx() {
|
||||||
|
let s = "Abcdef";
|
||||||
|
let lc: (usize, char) = s.char_indices().last().unwrap();
|
||||||
|
let t = &s[(lc.0 + lc.1.len_utf8())..];
|
||||||
|
dbg!(&t);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -201,7 +201,8 @@
|
||||||
},
|
},
|
||||||
"number_nd_tokens": {
|
"number_nd_tokens": {
|
||||||
"নাই": 0,
|
"নাই": 0,
|
||||||
"১": 1
|
"১": 1,
|
||||||
|
"১টা": 1
|
||||||
},
|
},
|
||||||
"album_types": {
|
"album_types": {
|
||||||
"ep": "Ep",
|
"ep": "Ep",
|
||||||
|
@ -2662,7 +2663,8 @@
|
||||||
"mrd": 9
|
"mrd": 9
|
||||||
},
|
},
|
||||||
"number_nd_tokens": {
|
"number_nd_tokens": {
|
||||||
"ingen": 0
|
"ingen": 0,
|
||||||
|
"én": 1
|
||||||
},
|
},
|
||||||
"album_types": {
|
"album_types": {
|
||||||
"album": "Album",
|
"album": "Album",
|
||||||
|
@ -2885,7 +2887,9 @@
|
||||||
"mi": 6,
|
"mi": 6,
|
||||||
"mil": 3
|
"mil": 3
|
||||||
},
|
},
|
||||||
"number_nd_tokens": {},
|
"number_nd_tokens": {
|
||||||
|
"um": 1
|
||||||
|
},
|
||||||
"album_types": {
|
"album_types": {
|
||||||
"audiolivro": "Audiobook",
|
"audiolivro": "Audiobook",
|
||||||
"ep": "Ep",
|
"ep": "Ep",
|
||||||
|
|
|
@ -16,7 +16,9 @@
|
||||||
"শঃ": null
|
"শঃ": null
|
||||||
},
|
},
|
||||||
"number_nd_tokens": {
|
"number_nd_tokens": {
|
||||||
"কোনো": null
|
"কোনো": null,
|
||||||
|
"ভিডিঅ’": null,
|
||||||
|
"১টা": 1
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"bn": {
|
"bn": {
|
||||||
|
@ -111,7 +113,8 @@
|
||||||
},
|
},
|
||||||
"no": {
|
"no": {
|
||||||
"number_nd_tokens": {
|
"number_nd_tokens": {
|
||||||
"avspillinger": null
|
"avspillinger": null,
|
||||||
|
"én": 1
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"or": {
|
"or": {
|
||||||
|
@ -129,6 +132,11 @@
|
||||||
"ਨੇ": null
|
"ਨੇ": null
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"pt": {
|
||||||
|
"number_nd_tokens": {
|
||||||
|
"um": 1
|
||||||
|
}
|
||||||
|
},
|
||||||
"ro": {
|
"ro": {
|
||||||
"number_nd_tokens": {
|
"number_nd_tokens": {
|
||||||
"abonat": null,
|
"abonat": null,
|
||||||
|
|
|
@ -1108,7 +1108,8 @@ fn search_empty(rp: RustyPipe) {
|
||||||
fn search_suggestion(rp: RustyPipe) {
|
fn search_suggestion(rp: RustyPipe) {
|
||||||
let result = tokio_test::block_on(rp.query().search_suggestion("hunger ga")).unwrap();
|
let result = tokio_test::block_on(rp.query().search_suggestion("hunger ga")).unwrap();
|
||||||
|
|
||||||
assert!(result.contains(&"hunger games".to_owned()));
|
assert!(result.iter().any(|s| s.starts_with("hunger games ")));
|
||||||
|
assert_gte(result.len(), 10, "search suggestions");
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
|
|
Loading…
Reference in a new issue