Merge pull request #283 from KotatsuApp/devi4
add sources and fix
commit
a61e441e79
@ -0,0 +1,179 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.all
|
||||||
|
|
||||||
|
import androidx.collection.ArraySet
|
||||||
|
import kotlinx.coroutines.async
|
||||||
|
import kotlinx.coroutines.awaitAll
|
||||||
|
import kotlinx.coroutines.coroutineScope
|
||||||
|
import org.jsoup.nodes.Element
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.PagedMangaParser
|
||||||
|
import org.koitharu.kotatsu.parsers.config.ConfigKey
|
||||||
|
import org.koitharu.kotatsu.parsers.model.*
|
||||||
|
import org.koitharu.kotatsu.parsers.util.*
|
||||||
|
import java.util.*
|
||||||
|
|
||||||
|
@MangaSourceParser("HENTAIFOX", "Hentai Fox", type = ContentType.HENTAI)
|
||||||
|
internal class HentaiFox(context: MangaLoaderContext) : PagedMangaParser(context, MangaSource.HENTAIFOX, 20) {
|
||||||
|
|
||||||
|
override val sortOrders: Set<SortOrder> = EnumSet.of(SortOrder.UPDATED)
|
||||||
|
override val configKeyDomain = ConfigKey.Domain("hentaifox.com")
|
||||||
|
|
||||||
|
override suspend fun getListPage(
|
||||||
|
page: Int,
|
||||||
|
query: String?,
|
||||||
|
tags: Set<MangaTag>?,
|
||||||
|
sortOrder: SortOrder,
|
||||||
|
): List<Manga> {
|
||||||
|
val tag = tags.oneOrThrowIfMany()
|
||||||
|
|
||||||
|
val url = buildString {
|
||||||
|
append("https://")
|
||||||
|
append(domain)
|
||||||
|
if (!tags.isNullOrEmpty()) {
|
||||||
|
append("/tag/")
|
||||||
|
append(tag?.key.orEmpty())
|
||||||
|
if (page > 1) {
|
||||||
|
append("/pag/")
|
||||||
|
append(page)
|
||||||
|
append("/")
|
||||||
|
}
|
||||||
|
} else if (!query.isNullOrEmpty()) {
|
||||||
|
append("/search/?q=")
|
||||||
|
append(query.urlEncoded())
|
||||||
|
if (page > 1) {
|
||||||
|
append("&page=")
|
||||||
|
append(page)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if (page > 2) {
|
||||||
|
append("/pag/")
|
||||||
|
append(page)
|
||||||
|
append("/")
|
||||||
|
} else if (page > 1) {
|
||||||
|
append("/page/")
|
||||||
|
append(page)
|
||||||
|
append("/")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
val doc = webClient.httpGet(url).parseHtml()
|
||||||
|
return doc.select(".lc_galleries .thumb").map { div ->
|
||||||
|
val href = div.selectFirstOrThrow(".inner_thumb a").attrAsRelativeUrl("href")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
title = div.select("h2.g_title").text(),
|
||||||
|
altTitle = null,
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(domain),
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
isNsfw = isNsfwSource,
|
||||||
|
coverUrl = div.selectFirstOrThrow("img").src().orEmpty(),
|
||||||
|
tags = emptySet(),
|
||||||
|
state = null,
|
||||||
|
author = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
//Tags are deliberately reduced because there are too many and this slows down the application.
|
||||||
|
//only the most popular ones are taken.
|
||||||
|
override suspend fun getTags(): Set<MangaTag> {
|
||||||
|
return coroutineScope {
|
||||||
|
(1..3).map { page ->
|
||||||
|
async { getTags(page) }
|
||||||
|
}
|
||||||
|
}.awaitAll().flattenTo(ArraySet(360))
|
||||||
|
}
|
||||||
|
|
||||||
|
private suspend fun getTags(page: Int): Set<MangaTag> {
|
||||||
|
val url = "https://$domain/tags/popular/pag/$page/"
|
||||||
|
val root = webClient.httpGet(url).parseHtml()
|
||||||
|
return root.parseTags()
|
||||||
|
}
|
||||||
|
|
||||||
|
private fun Element.parseTags() = select(".list_tags a.tag_btn").mapToSet {
|
||||||
|
val key = it.attr("href").removeSuffix('/').substringAfterLast('/')
|
||||||
|
MangaTag(
|
||||||
|
key = key,
|
||||||
|
title = it.selectFirstOrThrow("h3").text(),
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
override suspend fun getDetails(manga: Manga): Manga {
|
||||||
|
val doc = webClient.httpGet(manga.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val urlChapters = manga.url.replace("/gallery/", "/g/") + "1/"
|
||||||
|
return manga.copy(
|
||||||
|
altTitle = null,
|
||||||
|
tags = doc.select("ul.tags a.tag_btn ").mapNotNullToSet {
|
||||||
|
val key = it.attr("href").removeSuffix('/').substringAfterLast('/')
|
||||||
|
MangaTag(
|
||||||
|
key = key,
|
||||||
|
title = it.html().substringBefore("<span"),
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
},
|
||||||
|
author = doc.selectFirst("ul.artists a.tag_btn")?.html()?.substringBefore("<span"),
|
||||||
|
description = null,
|
||||||
|
chapters = listOf(
|
||||||
|
MangaChapter(
|
||||||
|
id = manga.id,
|
||||||
|
name = manga.title,
|
||||||
|
number = 1,
|
||||||
|
url = urlChapters,
|
||||||
|
scanlator = null,
|
||||||
|
uploadDate = 0,
|
||||||
|
branch = doc.selectFirstOrThrow("ul.languages a.tag_btn").html().substringBefore("<span"),
|
||||||
|
source = source,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getRelatedManga(seed: Manga): List<Manga> {
|
||||||
|
val doc = webClient.httpGet(seed.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val root = doc.body().selectFirstOrThrow(".related_galleries")
|
||||||
|
return root.select("div.thumb").mapNotNull { div ->
|
||||||
|
val a = div.selectFirst(".inner_thumb a") ?: return@mapNotNull null
|
||||||
|
val href = a.attrAsRelativeUrl("href")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(a.host ?: domain),
|
||||||
|
altTitle = null,
|
||||||
|
title = div.selectFirstOrThrow("h2.g_title").text(),
|
||||||
|
author = null,
|
||||||
|
coverUrl = div.selectFirst("img")?.src().orEmpty(),
|
||||||
|
tags = emptySet(),
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
state = null,
|
||||||
|
isNsfw = isNsfwSource,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
||||||
|
val doc = webClient.httpGet(chapter.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val totalPages = doc.selectFirstOrThrow(".total_pages").text().toInt()
|
||||||
|
val rawUrl = chapter.url.replace("/1/", "/")
|
||||||
|
return (1..totalPages).map {
|
||||||
|
val url = "$rawUrl$it/"
|
||||||
|
MangaPage(
|
||||||
|
id = generateUid(url),
|
||||||
|
url = url,
|
||||||
|
preview = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getPageUrl(page: MangaPage): String {
|
||||||
|
val doc = webClient.httpGet(page.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val root = doc.body()
|
||||||
|
return root.requireElementById("gimg").attrAsAbsoluteUrl("data-src")
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,180 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.all
|
||||||
|
|
||||||
|
import androidx.collection.ArraySet
|
||||||
|
import kotlinx.coroutines.async
|
||||||
|
import kotlinx.coroutines.awaitAll
|
||||||
|
import kotlinx.coroutines.coroutineScope
|
||||||
|
import org.jsoup.nodes.Element
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.PagedMangaParser
|
||||||
|
import org.koitharu.kotatsu.parsers.config.ConfigKey
|
||||||
|
import org.koitharu.kotatsu.parsers.model.*
|
||||||
|
import org.koitharu.kotatsu.parsers.util.*
|
||||||
|
import java.util.*
|
||||||
|
|
||||||
|
@MangaSourceParser("IMHENTAI", "ImHentai", type = ContentType.HENTAI)
|
||||||
|
internal class ImHentai(context: MangaLoaderContext) :
|
||||||
|
PagedMangaParser(context, MangaSource.IMHENTAI, pageSize = 20) {
|
||||||
|
|
||||||
|
override val sortOrders: Set<SortOrder> =
|
||||||
|
EnumSet.of(SortOrder.UPDATED, SortOrder.POPULARITY, SortOrder.RATING)
|
||||||
|
|
||||||
|
override val configKeyDomain = ConfigKey.Domain("imhentai.xxx")
|
||||||
|
|
||||||
|
override suspend fun getListPage(
|
||||||
|
page: Int,
|
||||||
|
query: String?,
|
||||||
|
tags: Set<MangaTag>?,
|
||||||
|
sortOrder: SortOrder,
|
||||||
|
): List<Manga> {
|
||||||
|
val tag = tags.oneOrThrowIfMany()
|
||||||
|
val url = buildString {
|
||||||
|
append("https://")
|
||||||
|
append(domain)
|
||||||
|
|
||||||
|
if (!query.isNullOrEmpty()) {
|
||||||
|
append("/search/?key=")
|
||||||
|
append(query.urlEncoded())
|
||||||
|
append("&page=")
|
||||||
|
append(page)
|
||||||
|
} else if (!tags.isNullOrEmpty()) {
|
||||||
|
append("/tag/")
|
||||||
|
append(tag?.key.orEmpty())
|
||||||
|
append("/")
|
||||||
|
when (sortOrder) {
|
||||||
|
SortOrder.UPDATED -> append("")
|
||||||
|
SortOrder.POPULARITY -> append("popular/")
|
||||||
|
else -> append("")
|
||||||
|
}
|
||||||
|
append("?page=")
|
||||||
|
append(page)
|
||||||
|
} else {
|
||||||
|
append("/search/?page=")
|
||||||
|
append(page)
|
||||||
|
when (sortOrder) {
|
||||||
|
SortOrder.UPDATED -> append("<=1&pp=0")
|
||||||
|
SortOrder.POPULARITY -> append("<=0&pp=1")
|
||||||
|
SortOrder.RATING -> append("<=0&pp=0")
|
||||||
|
else -> append("<=1&pp=0")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
val doc = webClient.httpGet(url).parseHtml()
|
||||||
|
return doc.select("div.galleries div.thumb").map { div ->
|
||||||
|
val a = div.selectFirstOrThrow(".inner_thumb a")
|
||||||
|
val href = a.attrAsRelativeUrl("href")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(domain),
|
||||||
|
coverUrl = a.selectFirst("img")?.src().orEmpty(),
|
||||||
|
title = div.selectFirst(".caption")?.text().orEmpty(),
|
||||||
|
altTitle = null,
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
tags = emptySet(),
|
||||||
|
author = null,
|
||||||
|
state = null,
|
||||||
|
source = source,
|
||||||
|
isNsfw = isNsfwSource,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
//Tags are deliberately reduced because there are too many and this slows down the application.
|
||||||
|
//only the most popular ones are taken.
|
||||||
|
override suspend fun getTags(): Set<MangaTag> {
|
||||||
|
return coroutineScope {
|
||||||
|
(1..3).map { page ->
|
||||||
|
async { getTags(page) }
|
||||||
|
}
|
||||||
|
}.awaitAll().flattenTo(ArraySet(360))
|
||||||
|
}
|
||||||
|
|
||||||
|
private suspend fun getTags(page: Int): Set<MangaTag> {
|
||||||
|
val url = "https://$domain/tags/popular/?page=$page"
|
||||||
|
val root = webClient.httpGet(url).parseHtml()
|
||||||
|
return root.parseTags()
|
||||||
|
}
|
||||||
|
|
||||||
|
private fun Element.parseTags() = select("div.stags a.tag_btn").mapToSet {
|
||||||
|
val href = it.attr("href").substringAfterLast("tag/").substringBeforeLast('/')
|
||||||
|
MangaTag(
|
||||||
|
key = href,
|
||||||
|
title = it.selectFirstOrThrow("h3.list_tag").text(),
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getDetails(manga: Manga): Manga = coroutineScope {
|
||||||
|
val fullUrl = manga.url.toAbsoluteUrl(domain)
|
||||||
|
val doc = webClient.httpGet(fullUrl).parseHtml()
|
||||||
|
manga.copy(
|
||||||
|
tags = doc.body().select("li:contains(Tags) a.tag").mapNotNullToSet {
|
||||||
|
val href = it.attr("href").substringAfterLast("tag/").substringBeforeLast('/')
|
||||||
|
val name = it.html().substringBeforeLast("<span")
|
||||||
|
MangaTag(
|
||||||
|
key = href,
|
||||||
|
title = name,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
},
|
||||||
|
author = doc.selectFirst("li:contains(Artists) a.tag")?.html()?.substringBefore("<span"),
|
||||||
|
chapters = listOf(
|
||||||
|
MangaChapter(
|
||||||
|
id = manga.id,
|
||||||
|
name = manga.title,
|
||||||
|
number = 1,
|
||||||
|
url = manga.url,
|
||||||
|
scanlator = null,
|
||||||
|
uploadDate = 0,
|
||||||
|
branch = doc.selectFirst("li:contains(Language) a.tag")?.html()?.substringBeforeLast("<span"),
|
||||||
|
source = source,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getRelatedManga(seed: Manga): List<Manga> {
|
||||||
|
val doc = webClient.httpGet(seed.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val root = doc.body().selectFirstOrThrow("div.related")
|
||||||
|
return root.select("div.thumb").mapNotNull { div ->
|
||||||
|
val a = div.selectFirstOrThrow(".inner_thumb a")
|
||||||
|
val href = a.attrAsRelativeUrl("href")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(domain),
|
||||||
|
coverUrl = a.selectFirst("img")?.src().orEmpty(),
|
||||||
|
title = div.selectFirst(".caption")?.text().orEmpty(),
|
||||||
|
altTitle = null,
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
tags = emptySet(),
|
||||||
|
author = null,
|
||||||
|
state = null,
|
||||||
|
source = source,
|
||||||
|
isNsfw = false,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
||||||
|
val fullUrl = chapter.url.toAbsoluteUrl(domain)
|
||||||
|
val doc = webClient.httpGet(fullUrl).parseHtml()
|
||||||
|
val totalPages = doc.selectFirstOrThrow(".pages").text().replace("Pages: ", "").toInt() + 1
|
||||||
|
val domainImg = doc.requireElementById("append_thumbs").selectFirstOrThrow("img").src()?.replace("1t.jpg", "")
|
||||||
|
val pages = ArrayList<MangaPage>(totalPages)
|
||||||
|
for (i in 1 until totalPages) {
|
||||||
|
val url = "$domainImg$i.jpg"
|
||||||
|
pages.add(
|
||||||
|
MangaPage(
|
||||||
|
id = generateUid(url),
|
||||||
|
url = url,
|
||||||
|
preview = null,
|
||||||
|
source = source,
|
||||||
|
),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
return pages
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -1,4 +1,4 @@
|
|||||||
package org.koitharu.kotatsu.parsers.site.en
|
package org.koitharu.kotatsu.parsers.site.all
|
||||||
|
|
||||||
import kotlinx.coroutines.async
|
import kotlinx.coroutines.async
|
||||||
import kotlinx.coroutines.coroutineScope
|
import kotlinx.coroutines.coroutineScope
|
||||||
@ -0,0 +1,167 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.ar
|
||||||
|
|
||||||
|
import kotlinx.coroutines.async
|
||||||
|
import kotlinx.coroutines.coroutineScope
|
||||||
|
import org.json.JSONArray
|
||||||
|
import org.json.JSONObject
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.PagedMangaParser
|
||||||
|
import org.koitharu.kotatsu.parsers.config.ConfigKey
|
||||||
|
import org.koitharu.kotatsu.parsers.model.*
|
||||||
|
import org.koitharu.kotatsu.parsers.util.*
|
||||||
|
import org.koitharu.kotatsu.parsers.util.json.mapJSON
|
||||||
|
import org.koitharu.kotatsu.parsers.util.json.mapJSONIndexed
|
||||||
|
import java.text.SimpleDateFormat
|
||||||
|
import java.util.*
|
||||||
|
|
||||||
|
@MangaSourceParser("FLIXSCANS", "Flix Scans", "ar")
|
||||||
|
internal class FlixScans(context: MangaLoaderContext) : PagedMangaParser(context, MangaSource.FLIXSCANS, 18) {
|
||||||
|
|
||||||
|
override val sortOrders: Set<SortOrder> = EnumSet.of(SortOrder.UPDATED)
|
||||||
|
override val configKeyDomain = ConfigKey.Domain("flixscans.com")
|
||||||
|
|
||||||
|
override suspend fun getListPage(
|
||||||
|
page: Int,
|
||||||
|
query: String?,
|
||||||
|
tags: Set<MangaTag>?,
|
||||||
|
sortOrder: SortOrder,
|
||||||
|
): List<Manga> {
|
||||||
|
val json = if (!query.isNullOrEmpty()) {
|
||||||
|
if (page > 1) {
|
||||||
|
return emptyList()
|
||||||
|
}
|
||||||
|
val url = "https://api.$domain/api/v1/search/serie"
|
||||||
|
val body = JSONObject()
|
||||||
|
body.put("title", query.urlEncoded())
|
||||||
|
webClient.httpPost(url, body).parseJson().getJSONArray("data")
|
||||||
|
} else if (!tags.isNullOrEmpty()) {
|
||||||
|
if (page > 1) {
|
||||||
|
return emptyList()
|
||||||
|
}
|
||||||
|
val tagQuery = tags.joinToString(separator = ",") { it.key }
|
||||||
|
val url = "https://api.$domain/api/v1/search/advance?=&genres=$tagQuery&serie_type=webtoon"
|
||||||
|
webClient.httpGet(url).parseJson().getJSONArray("data")
|
||||||
|
} else {
|
||||||
|
val url = "https://api.$domain/api/v1/webtoon/homepage/latest/home?page=$page"
|
||||||
|
webClient.httpGet(url).parseJson().getJSONArray("data")
|
||||||
|
}
|
||||||
|
return json.mapJSON { j ->
|
||||||
|
val href = "https://$domain/series/${j.getString("prefix")}-${j.getString("id")}-${j.getString("slug")}"
|
||||||
|
val cover = "https://api.$domain/storage/" + j.getString("thumbnail")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
title = j.getString("title"),
|
||||||
|
altTitle = null,
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(domain),
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
isNsfw = false,
|
||||||
|
coverUrl = cover,
|
||||||
|
tags = emptySet(),
|
||||||
|
state = when (j.getString("status")) {
|
||||||
|
"ongoing" -> MangaState.ONGOING
|
||||||
|
"completed" -> MangaState.FINISHED
|
||||||
|
else -> null
|
||||||
|
},
|
||||||
|
author = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getTags(): Set<MangaTag> {
|
||||||
|
val doc = webClient.httpGet("https://$domain/search/advance").parseHtml()
|
||||||
|
val json = JSONArray(doc.requireElementById("__NUXT_DATA__").data())
|
||||||
|
val tagsList = json.getJSONArray(3).toString().replace("[", "").replace("]", "").split(",")
|
||||||
|
return tagsList.mapNotNullToSet { idTag ->
|
||||||
|
val id = idTag.toInt()
|
||||||
|
val idKey = json.getJSONObject(id).getInt("id")
|
||||||
|
val key = json.get(idKey).toString()
|
||||||
|
val idName = json.getJSONObject(id).getInt("name")
|
||||||
|
val name = json.get(idName).toString()
|
||||||
|
MangaTag(
|
||||||
|
key = key,
|
||||||
|
title = name,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getDetails(manga: Manga): Manga = coroutineScope {
|
||||||
|
val doc = webClient.httpGet(manga.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val chaptersDeferred = async { loadChapters(manga.url) }
|
||||||
|
val json = JSONArray(doc.requireElementById("__NUXT_DATA__").data())
|
||||||
|
val descId = json.getJSONObject(6).getInt("story")
|
||||||
|
val desc = json.getString(descId)
|
||||||
|
val tagsId = json.getJSONObject(6).getInt("genres")
|
||||||
|
val tagsList = json.getJSONArray(tagsId).toString().replace("[", "").replace("]", "").split(",")
|
||||||
|
val ratingId = json.getJSONObject(6).getInt("rating")
|
||||||
|
val rating = json.getString(ratingId)
|
||||||
|
val nsfwId = json.getJSONObject(6).getInt("nsfw")
|
||||||
|
val nsfw = json.getBoolean(nsfwId)
|
||||||
|
manga.copy(
|
||||||
|
description = desc,
|
||||||
|
tags = tagsList.mapToSet { idTag ->
|
||||||
|
val id = idTag.toInt()
|
||||||
|
val idKey = json.getJSONObject(id).getInt("id")
|
||||||
|
val key = json.get(idKey).toString()
|
||||||
|
val idName = json.getJSONObject(id).getInt("name")
|
||||||
|
val name = json.get(idName).toString()
|
||||||
|
MangaTag(
|
||||||
|
key = key,
|
||||||
|
title = name,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
},
|
||||||
|
rating = rating?.toFloatOrNull()?.div(5f) ?: RATING_UNKNOWN,
|
||||||
|
isNsfw = nsfw,
|
||||||
|
chapters = chaptersDeferred.await(),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
private val dateFormat = SimpleDateFormat("yyyy-MM-dd", sourceLocale)
|
||||||
|
|
||||||
|
private suspend fun loadChapters(baseUrl: String): List<MangaChapter> {
|
||||||
|
val key = baseUrl.substringAfter("-").substringBefore("-")
|
||||||
|
val seriesKey = baseUrl.substringAfterLast("/").substringBefore("-")
|
||||||
|
val json = JSONArray(webClient.httpGet("https://api.$domain/api/v1/webtoon/chapters/$key-desc").parseRaw())
|
||||||
|
return json.mapJSONIndexed { i, j ->
|
||||||
|
val url = "https://$domain/read/webtoon/$seriesKey-${j.getString("id")}-${j.getString("slug")}"
|
||||||
|
val date = j.getString("createdAt").substringBeforeLast("T")
|
||||||
|
MangaChapter(
|
||||||
|
id = generateUid(url),
|
||||||
|
url = url,
|
||||||
|
name = j.getString("slug").replace('-', ' '),
|
||||||
|
number = i + 1,
|
||||||
|
branch = null,
|
||||||
|
uploadDate = dateFormat.tryParse(date),
|
||||||
|
scanlator = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
||||||
|
val fullUrl = chapter.url.toAbsoluteUrl(domain)
|
||||||
|
val doc = webClient.httpGet(fullUrl).parseHtml()
|
||||||
|
val json = JSONArray(doc.requireElementById("__NUXT_DATA__").data())
|
||||||
|
val chapterData = json.getJSONObject(6).getInt("chapterData")
|
||||||
|
val pageLocate = json.getJSONObject(chapterData).getInt("webtoon")
|
||||||
|
val jsonPages = json.getJSONArray(pageLocate)
|
||||||
|
val pages = ArrayList<MangaPage>(jsonPages.length())
|
||||||
|
for (i in 0 until jsonPages.length()) {
|
||||||
|
val id = jsonPages.getInt(i)
|
||||||
|
val url = "https://api.$domain/storage/" + json.getString(id)
|
||||||
|
pages.add(
|
||||||
|
MangaPage(
|
||||||
|
id = generateUid(url),
|
||||||
|
url = url,
|
||||||
|
preview = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
}
|
||||||
|
return pages
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,120 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.ar
|
||||||
|
|
||||||
|
import okhttp3.Headers
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.PagedMangaParser
|
||||||
|
import org.koitharu.kotatsu.parsers.config.ConfigKey
|
||||||
|
import org.koitharu.kotatsu.parsers.model.*
|
||||||
|
import org.koitharu.kotatsu.parsers.network.UserAgents
|
||||||
|
import org.koitharu.kotatsu.parsers.util.*
|
||||||
|
import java.util.*
|
||||||
|
|
||||||
|
@MangaSourceParser("MANGASTORM", "Manga Storm", "ar")
|
||||||
|
internal class MangaStorm(context: MangaLoaderContext) : PagedMangaParser(context, MangaSource.MANGASTORM, 30) {
|
||||||
|
|
||||||
|
override val sortOrders: Set<SortOrder> = EnumSet.of(SortOrder.POPULARITY)
|
||||||
|
override val configKeyDomain = ConfigKey.Domain("mangastorm.org")
|
||||||
|
|
||||||
|
override val headers: Headers = Headers.Builder()
|
||||||
|
.add("User-Agent", UserAgents.CHROME_DESKTOP)
|
||||||
|
.build()
|
||||||
|
|
||||||
|
override suspend fun getListPage(
|
||||||
|
page: Int,
|
||||||
|
query: String?,
|
||||||
|
tags: Set<MangaTag>?,
|
||||||
|
sortOrder: SortOrder,
|
||||||
|
): List<Manga> {
|
||||||
|
val tag = tags.oneOrThrowIfMany()
|
||||||
|
val url =
|
||||||
|
if (!tags.isNullOrEmpty()) {
|
||||||
|
buildString {
|
||||||
|
append("https://")
|
||||||
|
append(domain)
|
||||||
|
append("/categories/")
|
||||||
|
append(tag?.key.orEmpty())
|
||||||
|
append("?page=")
|
||||||
|
append(page)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
buildString {
|
||||||
|
append("https://")
|
||||||
|
append(domain)
|
||||||
|
append("/mangas?page=")
|
||||||
|
append(page)
|
||||||
|
if (!query.isNullOrEmpty()) {
|
||||||
|
append("&query=")
|
||||||
|
append(query.urlEncoded())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
val doc = webClient.httpGet(url).parseHtml()
|
||||||
|
return doc.select("div.row div.col").map { div ->
|
||||||
|
val href = div.selectFirstOrThrow("a").attrAsRelativeUrl("href")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
title = div.select(".manga-ct-title").text(),
|
||||||
|
altTitle = null,
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(domain),
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
isNsfw = false,
|
||||||
|
coverUrl = div.selectFirstOrThrow("img").src().orEmpty(),
|
||||||
|
tags = emptySet(),
|
||||||
|
state = null,
|
||||||
|
author = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getTags(): Set<MangaTag> = emptySet()
|
||||||
|
|
||||||
|
override suspend fun getDetails(manga: Manga): Manga {
|
||||||
|
val doc = webClient.httpGet(manga.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
|
||||||
|
val root = doc.selectFirstOrThrow(".card-body .col-lg-9")
|
||||||
|
|
||||||
|
return manga.copy(
|
||||||
|
altTitle = null,
|
||||||
|
state = null,
|
||||||
|
tags = root.select(".flex-wrap a").mapNotNullToSet { a ->
|
||||||
|
MangaTag(
|
||||||
|
key = a.attr("href").substringAfterLast('/'),
|
||||||
|
title = a.text(),
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
},
|
||||||
|
author = null,
|
||||||
|
description = root.selectFirstOrThrow(".card-text").text(),
|
||||||
|
chapters = doc.select(".card-body a.btn-fixed-width").mapChapters(reversed = true) { i, a ->
|
||||||
|
val url = a.attrAsRelativeUrl("href")
|
||||||
|
MangaChapter(
|
||||||
|
id = generateUid(url),
|
||||||
|
name = a.text(),
|
||||||
|
number = i + 1,
|
||||||
|
url = url,
|
||||||
|
scanlator = null,
|
||||||
|
uploadDate = 0,
|
||||||
|
branch = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
},
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
||||||
|
val fullUrl = chapter.url.toAbsoluteUrl(domain)
|
||||||
|
val doc = webClient.httpGet(fullUrl).parseHtml().requireElementById("content")
|
||||||
|
return doc.select("div.text-center .img-fluid").map { img ->
|
||||||
|
val url = img.src()?.toRelativeUrl(domain) ?: img.parseFailed("Image src not found")
|
||||||
|
MangaPage(
|
||||||
|
id = generateUid(url),
|
||||||
|
url = url,
|
||||||
|
preview = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,146 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.en
|
||||||
|
|
||||||
|
import kotlinx.coroutines.async
|
||||||
|
import kotlinx.coroutines.coroutineScope
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.PagedMangaParser
|
||||||
|
import org.koitharu.kotatsu.parsers.config.ConfigKey
|
||||||
|
import org.koitharu.kotatsu.parsers.model.*
|
||||||
|
import org.koitharu.kotatsu.parsers.util.*
|
||||||
|
import org.koitharu.kotatsu.parsers.util.json.mapJSONToSet
|
||||||
|
import java.util.*
|
||||||
|
|
||||||
|
@MangaSourceParser("FAKKU", "Fakku", "en", ContentType.HENTAI)
|
||||||
|
internal class Fakku(context: MangaLoaderContext) :
|
||||||
|
PagedMangaParser(context, MangaSource.FAKKU, pageSize = 25) {
|
||||||
|
|
||||||
|
override val sortOrders: Set<SortOrder> =
|
||||||
|
EnumSet.of(SortOrder.ALPHABETICAL, SortOrder.NEWEST, SortOrder.UPDATED)
|
||||||
|
|
||||||
|
override val configKeyDomain = ConfigKey.Domain("fakku.cc")
|
||||||
|
|
||||||
|
override suspend fun getListPage(
|
||||||
|
page: Int,
|
||||||
|
query: String?,
|
||||||
|
tags: Set<MangaTag>?,
|
||||||
|
sortOrder: SortOrder,
|
||||||
|
): List<Manga> {
|
||||||
|
val tag = tags.oneOrThrowIfMany()
|
||||||
|
val url = buildString {
|
||||||
|
append("https://")
|
||||||
|
append(domain)
|
||||||
|
when {
|
||||||
|
!query.isNullOrEmpty() -> {
|
||||||
|
append("/search?q=")
|
||||||
|
append(query.urlEncoded())
|
||||||
|
append("&")
|
||||||
|
}
|
||||||
|
|
||||||
|
!tags.isNullOrEmpty() -> {
|
||||||
|
append("/tags/")
|
||||||
|
append(tag?.key.orEmpty())
|
||||||
|
append("?")
|
||||||
|
}
|
||||||
|
|
||||||
|
else -> {
|
||||||
|
append("?")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
append("page=")
|
||||||
|
append(page)
|
||||||
|
append("&sort=")
|
||||||
|
when (sortOrder) {
|
||||||
|
SortOrder.ALPHABETICAL -> append("title")
|
||||||
|
SortOrder.NEWEST -> append("created_at")
|
||||||
|
SortOrder.UPDATED -> append("published_at")
|
||||||
|
else -> append("published_at")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
val doc = webClient.httpGet(url).parseHtml()
|
||||||
|
return doc.select("div.entries .entry a").map { a ->
|
||||||
|
val href = a.attrAsRelativeUrl("href")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(domain),
|
||||||
|
coverUrl = a.selectFirst("img")?.src().orEmpty(),
|
||||||
|
title = a.selectFirst(".title")?.text().orEmpty(),
|
||||||
|
altTitle = null,
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
tags = emptySet(),
|
||||||
|
author = null,
|
||||||
|
state = null,
|
||||||
|
source = source,
|
||||||
|
isNsfw = isNsfwSource,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getTags(): Set<MangaTag> {
|
||||||
|
val root = webClient.httpGet("https://$domain/tags").parseHtml()
|
||||||
|
return root.select("div.entries .entry a").mapToSet {
|
||||||
|
MangaTag(
|
||||||
|
key = it.attr("href").substringAfterLast("/"),
|
||||||
|
title = it.selectFirstOrThrow(".name").text(),
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getDetails(manga: Manga): Manga = coroutineScope {
|
||||||
|
val fullUrl = manga.url.toAbsoluteUrl(domain)
|
||||||
|
val doc = webClient.httpGet(fullUrl).parseHtml()
|
||||||
|
val genreDeferred = async {
|
||||||
|
webClient.httpGet(manga.url.toAbsoluteUrl(domain) + ".json").parseJson()
|
||||||
|
}
|
||||||
|
val genre = genreDeferred.await()
|
||||||
|
manga.copy(
|
||||||
|
author = doc.selectFirst("tr.artists a")?.text(),
|
||||||
|
tags = if (genre.toString().contains("tags")) {
|
||||||
|
genre.getJSONArray("tags").mapJSONToSet {
|
||||||
|
MangaTag(
|
||||||
|
key = it.getString("slug"),
|
||||||
|
title = it.getString("name"),
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
emptySet()
|
||||||
|
},
|
||||||
|
chapters = listOf(
|
||||||
|
MangaChapter(
|
||||||
|
id = manga.id,
|
||||||
|
name = manga.title,
|
||||||
|
number = 1,
|
||||||
|
url = manga.url + "/1",
|
||||||
|
scanlator = null,
|
||||||
|
uploadDate = 0,
|
||||||
|
branch = null,
|
||||||
|
source = source,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
||||||
|
val doc = webClient.httpGet(chapter.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val totalPages = doc.selectFirstOrThrow(".total").text().toInt()
|
||||||
|
val rawUrl = chapter.url.substringBeforeLast("/")
|
||||||
|
return (1..totalPages).map {
|
||||||
|
val url = "$rawUrl/$it"
|
||||||
|
MangaPage(
|
||||||
|
id = generateUid(url),
|
||||||
|
url = url,
|
||||||
|
preview = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getPageUrl(page: MangaPage): String {
|
||||||
|
val doc = webClient.httpGet(page.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val root = doc.body()
|
||||||
|
return root.selectFirstOrThrow(".page img").attrAsAbsoluteUrl("src")
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,168 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.en
|
||||||
|
|
||||||
|
import androidx.collection.ArraySet
|
||||||
|
import kotlinx.coroutines.async
|
||||||
|
import kotlinx.coroutines.awaitAll
|
||||||
|
import kotlinx.coroutines.coroutineScope
|
||||||
|
import org.jsoup.nodes.Element
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.PagedMangaParser
|
||||||
|
import org.koitharu.kotatsu.parsers.config.ConfigKey
|
||||||
|
import org.koitharu.kotatsu.parsers.model.*
|
||||||
|
import org.koitharu.kotatsu.parsers.util.*
|
||||||
|
import java.util.*
|
||||||
|
|
||||||
|
@MangaSourceParser("PURURIN", "Pururin", "en", ContentType.HENTAI)
|
||||||
|
internal class Pururin(context: MangaLoaderContext) :
|
||||||
|
PagedMangaParser(context, MangaSource.PURURIN, pageSize = 20) {
|
||||||
|
|
||||||
|
override val sortOrders: Set<SortOrder> =
|
||||||
|
EnumSet.of(SortOrder.UPDATED, SortOrder.POPULARITY, SortOrder.RATING, SortOrder.ALPHABETICAL)
|
||||||
|
|
||||||
|
override val configKeyDomain = ConfigKey.Domain("pururin.to")
|
||||||
|
|
||||||
|
override suspend fun getListPage(
|
||||||
|
page: Int,
|
||||||
|
query: String?,
|
||||||
|
tags: Set<MangaTag>?,
|
||||||
|
sortOrder: SortOrder,
|
||||||
|
): List<Manga> {
|
||||||
|
val tag = tags.oneOrThrowIfMany()
|
||||||
|
val url = buildString {
|
||||||
|
append("https://")
|
||||||
|
append(domain)
|
||||||
|
if (!query.isNullOrEmpty()) {
|
||||||
|
append("/search?q=")
|
||||||
|
append(query.urlEncoded())
|
||||||
|
append("&page=")
|
||||||
|
append(page)
|
||||||
|
} else {
|
||||||
|
append("/browse")
|
||||||
|
if (!tags.isNullOrEmpty()) {
|
||||||
|
append("/tags/content/")
|
||||||
|
append(tag?.key.orEmpty())
|
||||||
|
append("/")
|
||||||
|
}
|
||||||
|
append("?page=")
|
||||||
|
append(page)
|
||||||
|
append("&sort=")
|
||||||
|
when (sortOrder) {
|
||||||
|
SortOrder.UPDATED -> append("")
|
||||||
|
SortOrder.POPULARITY -> append("most-viewed")
|
||||||
|
SortOrder.RATING -> append("highest-rated")
|
||||||
|
SortOrder.ALPHABETICAL -> append("title")
|
||||||
|
else -> append("")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
val doc = webClient.httpGet(url).parseHtml()
|
||||||
|
return doc.select(".row-gallery a.card-gallery").map { a ->
|
||||||
|
val href = a.attrAsRelativeUrl("href")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(domain),
|
||||||
|
coverUrl = a.selectFirst("img.card-img-top")?.src().orEmpty(),
|
||||||
|
title = a.selectFirst(".title")?.text().orEmpty(),
|
||||||
|
altTitle = null,
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
tags = emptySet(),
|
||||||
|
author = null,
|
||||||
|
state = null,
|
||||||
|
source = source,
|
||||||
|
isNsfw = isNsfwSource,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getTags(): Set<MangaTag> {
|
||||||
|
return coroutineScope {
|
||||||
|
(1..4).map { page ->
|
||||||
|
async { getTags(page) }
|
||||||
|
}
|
||||||
|
}.awaitAll().flattenTo(ArraySet(360))
|
||||||
|
}
|
||||||
|
|
||||||
|
private suspend fun getTags(page: Int): Set<MangaTag> {
|
||||||
|
val url = "https://$domain/tags/content?order=uses&page=$page"
|
||||||
|
val root = webClient.httpGet(url).parseHtml()
|
||||||
|
return root.parseTags()
|
||||||
|
}
|
||||||
|
|
||||||
|
private fun Element.parseTags() = select("table tr td a").mapToSet {
|
||||||
|
val href = it.attr("href").substringAfterLast("content/").substringBeforeLast('/')
|
||||||
|
MangaTag(
|
||||||
|
key = href,
|
||||||
|
title = it.text(),
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getDetails(manga: Manga): Manga = coroutineScope {
|
||||||
|
val fullUrl = manga.url.toAbsoluteUrl(domain)
|
||||||
|
val doc = webClient.httpGet(fullUrl).parseHtml()
|
||||||
|
manga.copy(
|
||||||
|
description = doc.selectFirst("p.mb-2")?.text().orEmpty(),
|
||||||
|
rating = doc.selectFirst("td span.rating")?.attr("content")?.toFloatOrNull()?.div(5f) ?: RATING_UNKNOWN,
|
||||||
|
tags = doc.body().select("tr:contains(Contents) ul.list-inline a").mapNotNullToSet {
|
||||||
|
val href = it.attr("href").substringAfterLast("content/").substringBeforeLast('/')
|
||||||
|
MangaTag(
|
||||||
|
key = href,
|
||||||
|
title = it.text(),
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
},
|
||||||
|
author = doc.selectFirst("a[itemprop=author]")?.text(),
|
||||||
|
chapters = listOf(
|
||||||
|
MangaChapter(
|
||||||
|
id = manga.id,
|
||||||
|
name = manga.title,
|
||||||
|
number = 1,
|
||||||
|
url = manga.url,
|
||||||
|
scanlator = null,
|
||||||
|
uploadDate = 0,
|
||||||
|
branch = null,
|
||||||
|
source = source,
|
||||||
|
),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getRelatedManga(seed: Manga): List<Manga> {
|
||||||
|
val doc = webClient.httpGet(seed.url.toAbsoluteUrl(domain)).parseHtml()
|
||||||
|
val root = doc.body().selectFirstOrThrow(".row-gallery-small")
|
||||||
|
return root.select("a.card-gallery").mapNotNull { a ->
|
||||||
|
val href = a.attrAsRelativeUrl("href")
|
||||||
|
Manga(
|
||||||
|
id = generateUid(href),
|
||||||
|
url = href,
|
||||||
|
publicUrl = href.toAbsoluteUrl(domain),
|
||||||
|
coverUrl = a.selectFirst("img.card-img-top")?.src().orEmpty(),
|
||||||
|
title = a.selectFirst(".title")?.text().orEmpty(),
|
||||||
|
altTitle = null,
|
||||||
|
rating = RATING_UNKNOWN,
|
||||||
|
tags = emptySet(),
|
||||||
|
author = null,
|
||||||
|
state = null,
|
||||||
|
source = source,
|
||||||
|
isNsfw = false,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
||||||
|
val fullUrl = chapter.url.toAbsoluteUrl(domain)
|
||||||
|
val doc = webClient.httpGet(fullUrl).parseHtml()
|
||||||
|
return doc.select(".gallery-preview img").map { url ->
|
||||||
|
val img = url.src()?.toRelativeUrl(domain) ?: url.parseFailed("Image src not found")
|
||||||
|
val urlImage = img.replace("t.", ".")
|
||||||
|
MangaPage(
|
||||||
|
id = generateUid(urlImage),
|
||||||
|
url = urlImage,
|
||||||
|
preview = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,12 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.madara.ar
|
||||||
|
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.model.MangaSource
|
||||||
|
import org.koitharu.kotatsu.parsers.site.madara.MadaraParser
|
||||||
|
|
||||||
|
@MangaSourceParser("MANGATIME", "Manga Time", "ar")
|
||||||
|
internal class MangaTime(context: MangaLoaderContext) :
|
||||||
|
MadaraParser(context, MangaSource.MANGATIME, "mangatime.co") {
|
||||||
|
override val datePattern = "d MMMM، yyyy"
|
||||||
|
}
|
||||||
@ -0,0 +1,32 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.madara.en
|
||||||
|
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.model.ContentType
|
||||||
|
import org.koitharu.kotatsu.parsers.model.MangaChapter
|
||||||
|
import org.koitharu.kotatsu.parsers.model.MangaPage
|
||||||
|
import org.koitharu.kotatsu.parsers.model.MangaSource
|
||||||
|
import org.koitharu.kotatsu.parsers.site.madara.MadaraParser
|
||||||
|
import org.koitharu.kotatsu.parsers.util.*
|
||||||
|
|
||||||
|
@MangaSourceParser("MANHWADEN", "Manhwa Den", "en", ContentType.HENTAI)
|
||||||
|
internal class Manhwaden(context: MangaLoaderContext) :
|
||||||
|
MadaraParser(context, MangaSource.MANHWADEN, "www.manhwaden.com", 10) {
|
||||||
|
|
||||||
|
override val selectPage = "p img[src]"
|
||||||
|
|
||||||
|
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
||||||
|
val fullUrl = chapter.url.toAbsoluteUrl(domain)
|
||||||
|
val doc = webClient.httpGet(fullUrl).parseHtml()
|
||||||
|
val root = doc.body().selectFirstOrThrow(selectBodyPage)
|
||||||
|
return root.select(selectPage).map { img ->
|
||||||
|
val url = img.src()?.toRelativeUrl(domain) ?: img.parseFailed("Image src not found")
|
||||||
|
MangaPage(
|
||||||
|
id = generateUid(url),
|
||||||
|
url = url,
|
||||||
|
preview = null,
|
||||||
|
source = source,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@ -0,0 +1,12 @@
|
|||||||
|
package org.koitharu.kotatsu.parsers.site.madara.en
|
||||||
|
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||||
|
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||||
|
import org.koitharu.kotatsu.parsers.model.MangaSource
|
||||||
|
import org.koitharu.kotatsu.parsers.site.madara.MadaraParser
|
||||||
|
|
||||||
|
@MangaSourceParser("SECTSCANS", "Sect Scans", "en")
|
||||||
|
internal class SectScans(context: MangaLoaderContext) :
|
||||||
|
MadaraParser(context, MangaSource.SECTSCANS, "sectscans.com") {
|
||||||
|
override val listUrl = "comics/"
|
||||||
|
}
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue