|
|
|
|
@ -1,5 +1,6 @@
|
|
|
|
|
package org.koitharu.kotatsu.parsers.site.pt
|
|
|
|
|
|
|
|
|
|
import org.json.JSONObject
|
|
|
|
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
|
|
|
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
|
|
|
|
import org.koitharu.kotatsu.parsers.PagedMangaParser
|
|
|
|
|
@ -7,14 +8,16 @@ import org.koitharu.kotatsu.parsers.config.ConfigKey
|
|
|
|
|
import org.koitharu.kotatsu.parsers.model.*
|
|
|
|
|
import org.koitharu.kotatsu.parsers.util.*
|
|
|
|
|
import org.koitharu.kotatsu.parsers.util.json.mapJSON
|
|
|
|
|
import org.koitharu.kotatsu.parsers.util.json.mapJSONIndexed
|
|
|
|
|
import java.text.DateFormat
|
|
|
|
|
import java.text.SimpleDateFormat
|
|
|
|
|
import java.util.*
|
|
|
|
|
|
|
|
|
|
@MangaSourceParser("YUGENMANGAS", "YugenMangas.org", "pt")
|
|
|
|
|
@MangaSourceParser("YUGENMANGAS", "YugenMangas.net.br", "pt")
|
|
|
|
|
class YugenMangas(context: MangaLoaderContext) : PagedMangaParser(context, MangaSource.YUGENMANGAS, 28) {
|
|
|
|
|
|
|
|
|
|
override val sortOrders: Set<SortOrder> = EnumSet.of(SortOrder.ALPHABETICAL, SortOrder.UPDATED)
|
|
|
|
|
override val configKeyDomain = ConfigKey.Domain("yugenmangas.org")
|
|
|
|
|
override val sortOrders: Set<SortOrder> = EnumSet.of(SortOrder.ALPHABETICAL)
|
|
|
|
|
override val configKeyDomain = ConfigKey.Domain("yugenmangas.net.br")
|
|
|
|
|
|
|
|
|
|
override suspend fun getListPage(
|
|
|
|
|
page: Int,
|
|
|
|
|
@ -22,114 +25,114 @@ class YugenMangas(context: MangaLoaderContext) : PagedMangaParser(context, Manga
|
|
|
|
|
tags: Set<MangaTag>?,
|
|
|
|
|
sortOrder: SortOrder,
|
|
|
|
|
): List<Manga> {
|
|
|
|
|
if (!query.isNullOrEmpty()) {
|
|
|
|
|
if (page > 1) {
|
|
|
|
|
return emptyList()
|
|
|
|
|
}
|
|
|
|
|
val url = buildString {
|
|
|
|
|
append("https://")
|
|
|
|
|
append(domain)
|
|
|
|
|
append("/api/series/list/?query=")
|
|
|
|
|
append(query.urlEncoded())
|
|
|
|
|
}
|
|
|
|
|
val json = webClient.httpGet(url).parseJsonArray()
|
|
|
|
|
return json.mapJSON { j ->
|
|
|
|
|
val urlManga = "https://$domain/series/${j.getString("slug")}/"
|
|
|
|
|
Manga(
|
|
|
|
|
id = generateUid(urlManga),
|
|
|
|
|
url = urlManga,
|
|
|
|
|
publicUrl = urlManga,
|
|
|
|
|
title = j.getString("name"),
|
|
|
|
|
coverUrl = "",
|
|
|
|
|
altTitle = null,
|
|
|
|
|
rating = RATING_UNKNOWN,
|
|
|
|
|
tags = emptySet(),
|
|
|
|
|
description = null,
|
|
|
|
|
state = null,
|
|
|
|
|
author = null,
|
|
|
|
|
isNsfw = false,
|
|
|
|
|
source = source,
|
|
|
|
|
)
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
val url = buildString {
|
|
|
|
|
append("https://")
|
|
|
|
|
append(domain)
|
|
|
|
|
when (sortOrder) {
|
|
|
|
|
SortOrder.ALPHABETICAL -> append("/series")
|
|
|
|
|
SortOrder.UPDATED -> append("/updates")
|
|
|
|
|
else -> append("/updates")
|
|
|
|
|
val json =
|
|
|
|
|
if (!query.isNullOrEmpty()) {
|
|
|
|
|
if (page > 1) {
|
|
|
|
|
return emptyList()
|
|
|
|
|
}
|
|
|
|
|
val url = buildString {
|
|
|
|
|
append("https://api.")
|
|
|
|
|
append(domain)
|
|
|
|
|
append("/api/series/list/?query=")
|
|
|
|
|
append(query.urlEncoded())
|
|
|
|
|
}
|
|
|
|
|
webClient.httpGet(url).parseJsonArray()
|
|
|
|
|
} else {
|
|
|
|
|
if (page > 1) {
|
|
|
|
|
append("?page=")
|
|
|
|
|
append(page)
|
|
|
|
|
return emptyList()
|
|
|
|
|
}
|
|
|
|
|
val url = buildString {
|
|
|
|
|
append("https://api.")
|
|
|
|
|
append(domain)
|
|
|
|
|
append("/api/all_series/?page=1")
|
|
|
|
|
}
|
|
|
|
|
webClient.httpGet(url).parseJson().getJSONArray("series")
|
|
|
|
|
}
|
|
|
|
|
val doc = webClient.httpGet(url).parseHtml()
|
|
|
|
|
return doc.select(".gallery .mangas-gallery, .container-update-series .card-series-updates").map { div ->
|
|
|
|
|
val a = div.selectFirstOrThrow("a")
|
|
|
|
|
val href = a.attrAsRelativeUrl("href")
|
|
|
|
|
Manga(
|
|
|
|
|
id = generateUid(href),
|
|
|
|
|
url = href,
|
|
|
|
|
publicUrl = a.attrAsAbsoluteUrl("href"),
|
|
|
|
|
title = div.selectLastOrThrow(".title-serie, .name-manga").text(),
|
|
|
|
|
coverUrl = div.selectFirst("img")?.src().orEmpty(),
|
|
|
|
|
altTitle = null,
|
|
|
|
|
rating = RATING_UNKNOWN,
|
|
|
|
|
tags = emptySet(),
|
|
|
|
|
description = null,
|
|
|
|
|
state = null,
|
|
|
|
|
author = null,
|
|
|
|
|
isNsfw = false,
|
|
|
|
|
source = source,
|
|
|
|
|
)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return json.mapJSON { j ->
|
|
|
|
|
val slug = j.getString("slug")
|
|
|
|
|
Manga(
|
|
|
|
|
id = generateUid(slug),
|
|
|
|
|
url = slug,
|
|
|
|
|
publicUrl = slug,
|
|
|
|
|
title = j.getString("name"),
|
|
|
|
|
coverUrl = j.getString("cover"),
|
|
|
|
|
altTitle = null,
|
|
|
|
|
rating = RATING_UNKNOWN,
|
|
|
|
|
tags = emptySet(),
|
|
|
|
|
description = null,
|
|
|
|
|
state = null,
|
|
|
|
|
author = null,
|
|
|
|
|
isNsfw = isNsfwSource,
|
|
|
|
|
source = source,
|
|
|
|
|
)
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
override suspend fun getDetails(manga: Manga): Manga {
|
|
|
|
|
val doc = webClient.httpGet(manga.url.toAbsoluteUrl(domain)).parseHtml()
|
|
|
|
|
val dateFormat = SimpleDateFormat("dd.MM.yyyy", Locale.ROOT)
|
|
|
|
|
val chapters = doc.requireElementById("listadecapitulos")
|
|
|
|
|
val detailManga =
|
|
|
|
|
webClient.httpPost("https://api.$domain/api/serie_details/${manga.url}", emptyMap()).parseJson()
|
|
|
|
|
val body = JSONObject()
|
|
|
|
|
body.put("serie_slug", manga.url)
|
|
|
|
|
val chapterManga = webClient.httpPost("https://api.$domain/api/get_chapters_by_serie/", body).parseJson()
|
|
|
|
|
.getJSONArray("chapters")
|
|
|
|
|
val dateFormat = SimpleDateFormat("dd/MM/yyyy", sourceLocale)
|
|
|
|
|
return manga.copy(
|
|
|
|
|
description = doc.selectFirst(".sinopse .sinopse")?.html(),
|
|
|
|
|
author = doc.selectFirst(".author")?.text(),
|
|
|
|
|
coverUrl = doc.selectFirst(".side img")?.src().orEmpty(),
|
|
|
|
|
state = doc.selectFirst(".lancamento p")?.let {
|
|
|
|
|
when (it.text().lowercase()) {
|
|
|
|
|
description = detailManga.getString("synopsis"),
|
|
|
|
|
coverUrl = detailManga.getString("cover"),
|
|
|
|
|
altTitle = detailManga.getString("alternative_names"),
|
|
|
|
|
author = detailManga.getString("author"),
|
|
|
|
|
state = detailManga.getString("status")?.let {
|
|
|
|
|
when (it) {
|
|
|
|
|
"ongoing" -> MangaState.ONGOING
|
|
|
|
|
"completed", "finished" -> MangaState.FINISHED
|
|
|
|
|
"completed" -> MangaState.FINISHED
|
|
|
|
|
//"hiatus" -> MangaState.PAUSED
|
|
|
|
|
else -> null
|
|
|
|
|
}
|
|
|
|
|
},
|
|
|
|
|
chapters = chapters.select(".chapter").mapChapters(reversed = true) { i, div ->
|
|
|
|
|
val a = div.selectFirstOrThrow("a")
|
|
|
|
|
val href = a.attrAsRelativeUrl("href")
|
|
|
|
|
val title = div.selectFirstOrThrow(".chapter-title").text()
|
|
|
|
|
val dateText = div.selectFirstOrThrow(".chapter-lancado").text()
|
|
|
|
|
chapters = chapterManga.mapJSONIndexed { i, j ->
|
|
|
|
|
val url = "https://api.$domain/api/serie/${manga.url}/chapter/${j.getString("slug")}/images/imgs/"
|
|
|
|
|
MangaChapter(
|
|
|
|
|
id = generateUid(href),
|
|
|
|
|
name = title,
|
|
|
|
|
number = i + 1,
|
|
|
|
|
url = href,
|
|
|
|
|
id = generateUid(url),
|
|
|
|
|
name = j.getString("name"),
|
|
|
|
|
number = j.getString("name").removePrefix("Capítulo ").toInt(),
|
|
|
|
|
url = url,
|
|
|
|
|
scanlator = null,
|
|
|
|
|
uploadDate = dateFormat.tryParse(dateText),
|
|
|
|
|
uploadDate = parseChapterDate(
|
|
|
|
|
dateFormat,
|
|
|
|
|
j.getString("upload_date"),
|
|
|
|
|
),
|
|
|
|
|
branch = null,
|
|
|
|
|
source = source,
|
|
|
|
|
)
|
|
|
|
|
},
|
|
|
|
|
}.sortedBy { it.name },
|
|
|
|
|
)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
private fun parseChapterDate(dateFormat: DateFormat, date: String?): Long {
|
|
|
|
|
val d = date?.lowercase() ?: return 0
|
|
|
|
|
return when {
|
|
|
|
|
d.endsWith(" atrás") -> parseRelativeDate(date)
|
|
|
|
|
else -> dateFormat.tryParse(date)
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
private fun parseRelativeDate(date: String): Long {
|
|
|
|
|
val number = Regex("""(\d+)""").find(date)?.value?.toIntOrNull() ?: return 0
|
|
|
|
|
val cal = Calendar.getInstance()
|
|
|
|
|
|
|
|
|
|
return when {
|
|
|
|
|
WordSet("dia", "dias").anyWordIn(date) -> cal.apply { add(Calendar.DAY_OF_MONTH, -number) }.timeInMillis
|
|
|
|
|
WordSet("hora", "horas").anyWordIn(date) -> cal.apply { add(Calendar.HOUR, -number) }.timeInMillis
|
|
|
|
|
else -> 0
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
|
|
|
|
val fullUrl = chapter.url.toAbsoluteUrl(domain)
|
|
|
|
|
val apiUrl = webClient.httpGet(fullUrl).parseHtml().selectFirstOrThrow("script:containsData(imageUrls)").data()
|
|
|
|
|
.substringAfter("apiUrl = \"").substringBefore("\";")
|
|
|
|
|
val json = webClient.httpGet(apiUrl.toAbsoluteUrl(domain)).parseJson().getJSONArray("chapter_images")
|
|
|
|
|
val pages = ArrayList<MangaPage>(json.length())
|
|
|
|
|
for (i in 0 until json.length()) {
|
|
|
|
|
val img = "https://media.$domain/${json.getString(i)}"
|
|
|
|
|
val jsonPages = webClient.httpPost(chapter.url, emptyMap()).parseJson().getJSONArray("chapter_images")
|
|
|
|
|
val pages = ArrayList<MangaPage>(jsonPages.length())
|
|
|
|
|
for (i in 0 until jsonPages.length()) {
|
|
|
|
|
val img = "https://$domain/${jsonPages.getString(i)}"
|
|
|
|
|
pages.add(
|
|
|
|
|
MangaPage(
|
|
|
|
|
id = generateUid(img),
|
|
|
|
|
|