parent
50194df24d
commit
8d5fc945d4
@ -1,263 +0,0 @@
|
||||
package org.koitharu.kotatsu.parsers.site.en
|
||||
|
||||
import kotlinx.coroutines.Dispatchers
|
||||
import kotlinx.coroutines.delay
|
||||
import kotlinx.coroutines.withContext
|
||||
import okhttp3.Headers
|
||||
import okhttp3.Request
|
||||
import org.json.JSONArray
|
||||
import org.json.JSONObject
|
||||
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||
import org.koitharu.kotatsu.parsers.PagedMangaParser
|
||||
import org.koitharu.kotatsu.parsers.config.ConfigKey
|
||||
import org.koitharu.kotatsu.parsers.model.*
|
||||
import org.koitharu.kotatsu.parsers.util.*
|
||||
import org.koitharu.kotatsu.parsers.util.json.*
|
||||
import java.text.DateFormat
|
||||
import java.text.SimpleDateFormat
|
||||
import java.util.*
|
||||
|
||||
private const val MAX_RETRY_COUNT = 5
|
||||
|
||||
@MangaSourceParser("REAPERCOMICS", "ReaperComics", "en")
|
||||
internal class ReaperComics(context: MangaLoaderContext) :
|
||||
PagedMangaParser(context, MangaParserSource.REAPERCOMICS, pageSize = 20) {
|
||||
|
||||
override val availableSortOrders: Set<SortOrder> = EnumSet.of(
|
||||
SortOrder.UPDATED,
|
||||
SortOrder.ALPHABETICAL,
|
||||
SortOrder.POPULARITY,
|
||||
SortOrder.NEWEST,
|
||||
SortOrder.ALPHABETICAL_DESC,
|
||||
)
|
||||
|
||||
override val configKeyDomain = ConfigKey.Domain("reaperscans.com")
|
||||
|
||||
private val userAgentKey = ConfigKey.UserAgent(context.getDefaultUserAgent())
|
||||
|
||||
private val baseHeaders: Headers
|
||||
get() = Headers.Builder().add("User-Agent", config[userAgentKey]).build()
|
||||
|
||||
override val headers
|
||||
get() = getApiHeaders()
|
||||
|
||||
private fun getApiHeaders(): Headers {
|
||||
val userCookie = context.cookieJar.getCookies(domain).find {
|
||||
it.name == "user"
|
||||
} ?: return baseHeaders
|
||||
val jo = JSONObject(userCookie.value.urlDecode())
|
||||
val accessToken = jo.getStringOrNull("access_token") ?: return baseHeaders
|
||||
return baseHeaders.newBuilder().add("authorization", "bearer $accessToken").build()
|
||||
}
|
||||
|
||||
override suspend fun getListPage(page: Int, filter: MangaListFilter?): List<Manga> {
|
||||
if (page > 1) return emptyList()
|
||||
|
||||
val url = buildString {
|
||||
append("https://")
|
||||
append("api.$domain")
|
||||
append("/query?page=$page&perPage=9999&series_type=Comic")
|
||||
when (filter) {
|
||||
is MangaListFilter.Search -> {
|
||||
append("&query_string=")
|
||||
append(filter.query.urlEncoded())
|
||||
}
|
||||
|
||||
is MangaListFilter.Advanced -> {
|
||||
append("&orderBy=")
|
||||
val order = when (filter.sortOrder) {
|
||||
SortOrder.UPDATED -> "updated_at"
|
||||
SortOrder.POPULARITY -> "total_views"
|
||||
SortOrder.NEWEST -> "created_at"
|
||||
SortOrder.ALPHABETICAL -> "title"
|
||||
SortOrder.ALPHABETICAL_DESC -> "title"
|
||||
else -> "updated_at"
|
||||
}
|
||||
append(order)
|
||||
val sortOrder = if (filter.sortOrder == SortOrder.ALPHABETICAL_DESC) "desc" else "asc"
|
||||
append("&order=$sortOrder")
|
||||
|
||||
filter.states.oneOrThrowIfMany()?.let {
|
||||
append("&status=")
|
||||
append(
|
||||
when (it) {
|
||||
MangaState.ONGOING -> "Ongoing"
|
||||
MangaState.FINISHED -> "Completed"
|
||||
MangaState.ABANDONED -> "Dropped"
|
||||
MangaState.PAUSED -> "Hiatus"
|
||||
else -> "All"
|
||||
},
|
||||
)
|
||||
}
|
||||
if (filter.tags.isNotEmpty()) {
|
||||
append("&tags_ids=")
|
||||
append(filter.tags.joinToString(separator = "%") { it.key })
|
||||
}
|
||||
}
|
||||
|
||||
null -> {
|
||||
append("&orderBy=updated_at")
|
||||
append("&order=asc")
|
||||
append("&adult=true")
|
||||
append("&status=All")
|
||||
}
|
||||
}
|
||||
}
|
||||
return parseMangaList(webClient.httpGet(url).parseJson())
|
||||
}
|
||||
|
||||
|
||||
private fun parseMangaList(response: JSONObject): List<Manga> {
|
||||
return response.getJSONArray("data").mapJSON { it ->
|
||||
val id = it.getLong("id")
|
||||
val url = "/comic/${it.getString("series_slug")}"
|
||||
val title = it.getString("title")
|
||||
val thumbnailPath = it.getString("thumbnail")
|
||||
Manga(
|
||||
id = id,
|
||||
url = url,
|
||||
title = title,
|
||||
altTitle = it.getString("alternative_names").takeIf { it.isNotBlank() },
|
||||
publicUrl = url.toAbsoluteUrl(domain),
|
||||
description = it.getString("description"),
|
||||
rating = it.getFloatOrDefault("rating", RATING_UNKNOWN) / 5f,
|
||||
isNsfw = isNsfwSource,
|
||||
coverUrl = "https://media.reaperscans.com/file/4SRBHm//$thumbnailPath",
|
||||
tags = emptySet(),
|
||||
state = when (it.getString("status")) {
|
||||
"Ongoing" -> MangaState.ONGOING
|
||||
"Completed" -> MangaState.FINISHED
|
||||
"Dropped" -> MangaState.ABANDONED
|
||||
"Hiatus" -> MangaState.PAUSED
|
||||
else -> null
|
||||
},
|
||||
author = null,
|
||||
source = source,
|
||||
)
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
override suspend fun getAvailableTags(): Set<MangaTag> {
|
||||
val doc = webClient.httpGet("https://$domain/comics").parseHtml()
|
||||
val scriptContent = doc.select("script").find {
|
||||
it.data().contains("tags")
|
||||
}?.data()
|
||||
|
||||
if (scriptContent != null) {
|
||||
val jsonString = scriptContent.substringAfter("push(").substringBeforeLast(")")
|
||||
val jsonArray = JSONArray(jsonString)
|
||||
val childrenArray = jsonArray.getString(1)
|
||||
val tagsString = childrenArray.substringAfter("tags:[").substringBeforeLast("]")
|
||||
val tagObjects = tagsString.split("},{")
|
||||
|
||||
return tagObjects.mapNotNullTo(mutableSetOf()) { tagString ->
|
||||
|
||||
val id = tagString.substringAfter("\"id\":").substringBefore(",")
|
||||
val name = tagString.substringAfter("\"name\":\"").substringBefore("\"")
|
||||
if (id.isNotEmpty() && name.isNotEmpty()) {
|
||||
MangaTag(
|
||||
key = id,
|
||||
title = name.toTitleCase(sourceLocale),
|
||||
source = source,
|
||||
)
|
||||
} else {
|
||||
null
|
||||
}
|
||||
}
|
||||
}
|
||||
return emptySet()
|
||||
}
|
||||
|
||||
override fun onCreateConfig(keys: MutableCollection<ConfigKey<*>>) {
|
||||
super.onCreateConfig(keys)
|
||||
keys.add(userAgentKey)
|
||||
}
|
||||
|
||||
private val dateFormat = SimpleDateFormat("yyyy-MM-dd'T'HH:mm:ss.SSS'Z'", sourceLocale)
|
||||
|
||||
override suspend fun getDetails(manga: Manga): Manga {
|
||||
val seriesid = manga.id
|
||||
val url = "https://api.$domain/chapter/query?page=1&perPage=9999&series_id=$seriesid"
|
||||
val response = makeRequest(url)
|
||||
val data = response.getJSONArray("data")
|
||||
return manga.copy(
|
||||
chapters = data.mapJSONIndexed { index, it ->
|
||||
val chapterUrl =
|
||||
"/series/${it.getJSONObject("series").getString("series_slug")}/${it.getString("chapter_slug")}"
|
||||
MangaChapter(
|
||||
id = it.getLong("id"),
|
||||
name = it.getString("chapter_name"),
|
||||
number = (data.length() - index).toFloat(),
|
||||
volume = 0,
|
||||
url = chapterUrl,
|
||||
scanlator = null,
|
||||
uploadDate = parseChapterDate(dateFormat, it.getString("created_at")),
|
||||
branch = null,
|
||||
source = source,
|
||||
)
|
||||
},
|
||||
)
|
||||
}
|
||||
|
||||
private fun parseChapterDate(dateFormat: DateFormat, date: String?): Long {
|
||||
return try {
|
||||
dateFormat.tryParse(date)
|
||||
} catch (e: Exception) {
|
||||
0L
|
||||
}
|
||||
}
|
||||
|
||||
private val pageSelector = "div#content div.container img"
|
||||
|
||||
override suspend fun getPages(chapter: MangaChapter): List<MangaPage> {
|
||||
val doc = webClient.httpGet(chapter.url.toAbsoluteUrl(domain)).parseHtml()
|
||||
val processedUrls = mutableSetOf<String>()
|
||||
|
||||
return doc.select(pageSelector).mapNotNull { img ->
|
||||
val url = img.attr("data-cfsrc").takeIf { it.isNotBlank() }
|
||||
?: img.attr("src").takeIf { it.isNotBlank() }
|
||||
?: img.selectFirst("noscript img")?.attr("src")
|
||||
?: return@mapNotNull null
|
||||
|
||||
val relativeUrl = url.toRelativeUrl(domain)
|
||||
|
||||
if (relativeUrl !in processedUrls) {
|
||||
processedUrls.add(relativeUrl)
|
||||
MangaPage(
|
||||
id = generateUid(relativeUrl),
|
||||
url = relativeUrl,
|
||||
preview = null,
|
||||
source = source,
|
||||
)
|
||||
} else {
|
||||
null
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private suspend fun makeRequest(url: String): JSONObject {
|
||||
var retryCount = 0
|
||||
val backoffDelay = 2000L // Initial delay (milliseconds)
|
||||
val request = Request.Builder().url(url).headers(headers).build()
|
||||
|
||||
while (true) {
|
||||
try {
|
||||
val response = context.httpClient.newCall(request).execute().parseJson()
|
||||
return response
|
||||
|
||||
} catch (e: Exception) {
|
||||
// Log or handle the exception as needed
|
||||
if (++retryCount <= MAX_RETRY_COUNT) {
|
||||
withContext(Dispatchers.Default) {
|
||||
delay(backoffDelay)
|
||||
}
|
||||
} else {
|
||||
throw e
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@ -1,149 +1,10 @@
|
||||
package org.koitharu.kotatsu.parsers.site.heancms.en
|
||||
|
||||
import org.json.JSONArray
|
||||
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||
import org.koitharu.kotatsu.parsers.model.*
|
||||
import org.koitharu.kotatsu.parsers.site.heancms.HeanCms
|
||||
import org.koitharu.kotatsu.parsers.util.*
|
||||
import org.koitharu.kotatsu.parsers.util.json.getFloatOrDefault
|
||||
import org.koitharu.kotatsu.parsers.util.json.getStringOrNull
|
||||
import org.koitharu.kotatsu.parsers.util.json.mapJSON
|
||||
import org.koitharu.kotatsu.parsers.util.json.unescapeJson
|
||||
import java.text.SimpleDateFormat
|
||||
import java.util.*
|
||||
|
||||
@MangaSourceParser("OMEGASCANS", "OmegaScans", "en", ContentType.HENTAI)
|
||||
internal class OmegaScans(context: MangaLoaderContext) :
|
||||
HeanCms(context, MangaParserSource.OMEGASCANS, "omegascans.org") {
|
||||
override suspend fun getListPage(page: Int, filter: MangaListFilter?): List<Manga> {
|
||||
val url = buildString {
|
||||
append("https://api.")
|
||||
append(domain)
|
||||
append("/query?adult=true&query_string=")
|
||||
when (filter) {
|
||||
is MangaListFilter.Search -> {
|
||||
append(filter.query.urlEncoded())
|
||||
}
|
||||
|
||||
is MangaListFilter.Advanced -> {
|
||||
|
||||
filter.states.oneOrThrowIfMany()?.let {
|
||||
append("&series_status=")
|
||||
append(
|
||||
when (it) {
|
||||
MangaState.ONGOING -> "Ongoing"
|
||||
MangaState.FINISHED -> "Completed"
|
||||
MangaState.ABANDONED -> "Dropped"
|
||||
MangaState.PAUSED -> "Hiatus"
|
||||
else -> ""
|
||||
},
|
||||
)
|
||||
|
||||
}
|
||||
append("&orderBy=")
|
||||
when (filter.sortOrder) {
|
||||
SortOrder.POPULARITY -> append("total_views&order=desc")
|
||||
SortOrder.UPDATED -> append("latest&order=desc")
|
||||
SortOrder.NEWEST -> append("created_at&order=desc")
|
||||
SortOrder.ALPHABETICAL -> append("title&order=asc")
|
||||
SortOrder.ALPHABETICAL_DESC -> append("title&order=desc")
|
||||
else -> append("latest&order=desc")
|
||||
}
|
||||
append("&series_type=All&perPage=$pageSize")
|
||||
append("&tags_ids=")
|
||||
append("[".urlEncoded())
|
||||
append(filter.tags.joinToString(",") { it.key })
|
||||
append("]".urlEncoded())
|
||||
|
||||
}
|
||||
|
||||
null -> {}
|
||||
}
|
||||
append("&page=")
|
||||
append(page.toString())
|
||||
}
|
||||
val json = webClient.httpGet(url).parseJson()
|
||||
return json.getJSONArray("data").mapJSON { j ->
|
||||
val slug = j.getString("series_slug")
|
||||
val urlManga = "https://$domain/$pathManga/$slug"
|
||||
val cover = if (j.getString("thumbnail").contains('/')) {
|
||||
j.getString("thumbnail")
|
||||
} else {
|
||||
"https://api.$domain/${j.getString("thumbnail")}"
|
||||
}
|
||||
Manga(
|
||||
id = j.getLong("id"),
|
||||
title = j.getString("title"),
|
||||
altTitle = null,
|
||||
url = urlManga.toRelativeUrl(domain),
|
||||
publicUrl = urlManga,
|
||||
rating = j.getFloatOrDefault("rating", RATING_UNKNOWN) / 5f,
|
||||
isNsfw = true,
|
||||
coverUrl = cover,
|
||||
tags = setOf(),
|
||||
state = when (j.getString("status")) {
|
||||
"Ongoing" -> MangaState.ONGOING
|
||||
"Completed" -> MangaState.FINISHED
|
||||
"Dropped" -> MangaState.ABANDONED
|
||||
"Hiatus" -> MangaState.PAUSED
|
||||
else -> null
|
||||
},
|
||||
author = j.getStringOrNull("author"),
|
||||
source = source,
|
||||
description = j.getString("description"),
|
||||
)
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
override suspend fun getDetails(manga: Manga): Manga {
|
||||
val url = buildString {
|
||||
append("https://api.")
|
||||
append(domain)
|
||||
append("/chapter/query?perPage=9999&series_id=" + manga.id)
|
||||
}
|
||||
val json = webClient.httpGet(url).parseJson()
|
||||
val dateFormat = SimpleDateFormat(datePattern, Locale.ENGLISH)
|
||||
|
||||
val chaptersJsonArray = json.getJSONArray("data")
|
||||
var totalChapters = json.getJSONObject("meta").getInt("total").toFloat()
|
||||
val chapters = chaptersJsonArray.mapJSON { j ->
|
||||
val slug = j.getJSONObject("series").getString("series_slug")
|
||||
val chapterUrl = "https://$domain/$pathManga/$slug/${j.getString("chapter_slug")}"
|
||||
val date = j.getString("created_at").substringBeforeLast("T")
|
||||
MangaChapter(
|
||||
id = j.getLong("id"),
|
||||
url = chapterUrl,
|
||||
name = j.getString("chapter_name"),
|
||||
number = totalChapters--,
|
||||
volume = 0,
|
||||
branch = null,
|
||||
uploadDate = dateFormat.tryParse(date),
|
||||
scanlator = null,
|
||||
source = source,
|
||||
)
|
||||
}
|
||||
|
||||
return manga.copy(
|
||||
chapters = chapters.reversed(),
|
||||
)
|
||||
}
|
||||
|
||||
override suspend fun getAvailableTags(): Set<MangaTag> {
|
||||
val doc = webClient.httpGet("https://$domain/comics").parseHtml()
|
||||
|
||||
val regex = Regex("\"tags\\\\.*?(\\[.+?])")
|
||||
val tags = doc.select("script").firstNotNullOf { script ->
|
||||
regex.find(script.html())?.groupValues?.getOrNull(1)
|
||||
}.unescapeJson()
|
||||
return JSONArray(tags).mapJSON {
|
||||
MangaTag(
|
||||
key = it.getInt("id").toString(),
|
||||
title = it.getString("name").toTitleCase(sourceLocale),
|
||||
source = source,
|
||||
)
|
||||
}.toSet()
|
||||
}
|
||||
|
||||
}
|
||||
HeanCms(context, MangaParserSource.OMEGASCANS, "omegascans.org")
|
||||
|
||||
@ -0,0 +1,13 @@
|
||||
package org.koitharu.kotatsu.parsers.site.heancms.en
|
||||
|
||||
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||
import org.koitharu.kotatsu.parsers.model.*
|
||||
import org.koitharu.kotatsu.parsers.site.heancms.HeanCms
|
||||
|
||||
@MangaSourceParser("REAPERCOMICS", "ReaperComics", "en")
|
||||
internal class ReaperComics(context: MangaLoaderContext) :
|
||||
HeanCms(context, MangaParserSource.REAPERCOMICS, "reaperscans.com") {
|
||||
override val cdn = "media.reaperscans.com/file/4SRBHm//"
|
||||
override val paramsUpdated = "updated_at"
|
||||
}
|
||||
@ -1,148 +1,11 @@
|
||||
package org.koitharu.kotatsu.parsers.site.heancms.fr
|
||||
|
||||
import org.json.JSONArray
|
||||
|
||||
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||
import org.koitharu.kotatsu.parsers.model.*
|
||||
import org.koitharu.kotatsu.parsers.site.heancms.HeanCms
|
||||
import org.koitharu.kotatsu.parsers.util.*
|
||||
import org.koitharu.kotatsu.parsers.util.json.*
|
||||
import java.text.SimpleDateFormat
|
||||
import java.util.*
|
||||
|
||||
@MangaSourceParser("PERF_SCAN", "PerfScan", "fr")
|
||||
internal class PerfScan(context: MangaLoaderContext) :
|
||||
HeanCms(context, MangaParserSource.PERF_SCAN, "perf-scan.fr") {
|
||||
|
||||
override suspend fun getListPage(page: Int, filter: MangaListFilter?): List<Manga> {
|
||||
val url = buildString {
|
||||
append("https://api.")
|
||||
append(domain)
|
||||
append("/query?adult=true&query_string=")
|
||||
when (filter) {
|
||||
is MangaListFilter.Search -> {
|
||||
append(filter.query.urlEncoded())
|
||||
}
|
||||
|
||||
is MangaListFilter.Advanced -> {
|
||||
|
||||
filter.states.oneOrThrowIfMany()?.let {
|
||||
append("&status=")
|
||||
append(
|
||||
when (it) {
|
||||
MangaState.ONGOING -> "Ongoing"
|
||||
MangaState.FINISHED -> "Completed"
|
||||
MangaState.ABANDONED -> "Dropped"
|
||||
MangaState.PAUSED -> "Hiatus"
|
||||
else -> ""
|
||||
},
|
||||
)
|
||||
|
||||
}
|
||||
append("&orderBy=")
|
||||
when (filter.sortOrder) {
|
||||
SortOrder.POPULARITY -> append("total_views&order=desc")
|
||||
SortOrder.UPDATED -> append("latest&order=desc")
|
||||
SortOrder.NEWEST -> append("created_at&order=desc")
|
||||
SortOrder.ALPHABETICAL -> append("title&order=asc")
|
||||
SortOrder.ALPHABETICAL_DESC -> append("title&order=desc")
|
||||
else -> append("latest&order=desc")
|
||||
}
|
||||
append("&series_type=All&perPage=")
|
||||
append(pageSize)
|
||||
append("&tags_ids=")
|
||||
append("[".urlEncoded())
|
||||
filter.tags.joinTo(this, ",") { it.key }
|
||||
append("]".urlEncoded())
|
||||
|
||||
}
|
||||
|
||||
null -> {}
|
||||
}
|
||||
append("&page=")
|
||||
append(page)
|
||||
}
|
||||
val json = webClient.httpGet(url).parseJson()
|
||||
return json.getJSONArray("data").mapJSON { j ->
|
||||
val slug = j.getString("series_slug")
|
||||
val urlManga = "https://$domain/$pathManga/$slug"
|
||||
val cover = if (j.getString("thumbnail").contains('/')) {
|
||||
j.getString("thumbnail")
|
||||
} else {
|
||||
"https://api.$domain/${j.getString("thumbnail")}"
|
||||
}
|
||||
Manga(
|
||||
id = j.getLong("id"),
|
||||
title = j.getString("title"),
|
||||
altTitle = null,
|
||||
url = urlManga.toRelativeUrl(domain),
|
||||
publicUrl = urlManga,
|
||||
rating = j.getFloatOrDefault("rating", RATING_UNKNOWN) / 5f,
|
||||
isNsfw = isNsfwSource,
|
||||
coverUrl = cover,
|
||||
tags = setOf(),
|
||||
state = when (j.getStringOrNull("status")) {
|
||||
"Ongoing" -> MangaState.ONGOING
|
||||
"Completed" -> MangaState.FINISHED
|
||||
"Dropped" -> MangaState.ABANDONED
|
||||
"Hiatus" -> MangaState.PAUSED
|
||||
else -> null
|
||||
},
|
||||
author = j.getStringOrNull("author"),
|
||||
source = source,
|
||||
description = j.getString("description"),
|
||||
)
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
override suspend fun getDetails(manga: Manga): Manga {
|
||||
val url = buildString {
|
||||
append("https://api.")
|
||||
append(domain)
|
||||
append("/chapter/query?perPage=9999&series_id=")
|
||||
append(manga.id)
|
||||
}
|
||||
val json = webClient.httpGet(url).parseJson()
|
||||
val dateFormat = SimpleDateFormat(datePattern, Locale.ENGLISH)
|
||||
|
||||
val chaptersJsonArray = json.getJSONArray("data")
|
||||
var totalChapters = json.getJSONObject("meta").getInt("total").toFloat()
|
||||
val chapters = chaptersJsonArray.mapJSON { j ->
|
||||
val slug = j.getJSONObject("series").getString("series_slug")
|
||||
val chapterUrl = "https://$domain/$pathManga/$slug/${j.getString("chapter_slug")}"
|
||||
val date = j.getString("created_at").substringBeforeLast("T")
|
||||
MangaChapter(
|
||||
id = j.getLong("id"),
|
||||
url = chapterUrl,
|
||||
name = j.getString("chapter_name"),
|
||||
number = totalChapters--,
|
||||
volume = 0,
|
||||
branch = null,
|
||||
uploadDate = dateFormat.tryParse(date),
|
||||
scanlator = null,
|
||||
source = source,
|
||||
)
|
||||
}
|
||||
|
||||
return manga.copy(
|
||||
chapters = chapters.reversed(),
|
||||
)
|
||||
}
|
||||
|
||||
override suspend fun getAvailableTags(): Set<MangaTag> {
|
||||
val doc = webClient.httpGet("https://$domain/comics").parseHtml()
|
||||
|
||||
val regex = Regex("\"tags\\\\.*?(\\[.+?])")
|
||||
val tags = doc.select("script").firstNotNullOf { script ->
|
||||
regex.find(script.html())?.groupValues?.getOrNull(1)
|
||||
}.unescapeJson()
|
||||
return JSONArray(tags).mapJSONToSet {
|
||||
MangaTag(
|
||||
key = it.getInt("id").toString(),
|
||||
title = it.getString("name").toTitleCase(sourceLocale),
|
||||
source = source,
|
||||
)
|
||||
}
|
||||
}
|
||||
}
|
||||
HeanCms(context, MangaParserSource.PERF_SCAN, "perf-scan.fr")
|
||||
|
||||
@ -1,150 +1,13 @@
|
||||
package org.koitharu.kotatsu.parsers.site.heancms.pt
|
||||
|
||||
import org.json.JSONArray
|
||||
import org.koitharu.kotatsu.parsers.MangaLoaderContext
|
||||
import org.koitharu.kotatsu.parsers.MangaSourceParser
|
||||
import org.koitharu.kotatsu.parsers.model.*
|
||||
import org.koitharu.kotatsu.parsers.site.heancms.HeanCms
|
||||
import org.koitharu.kotatsu.parsers.util.*
|
||||
import org.koitharu.kotatsu.parsers.util.json.*
|
||||
import java.text.SimpleDateFormat
|
||||
import java.util.*
|
||||
|
||||
@MangaSourceParser("MODESCANLATOR", "ModeScanlator", "pt")
|
||||
internal class ModeScanlator(
|
||||
context: MangaLoaderContext,
|
||||
) : HeanCms(context, MangaParserSource.MODESCANLATOR, "site.modescanlator.net") {
|
||||
|
||||
private val domainNoSite = domain.removePrefix("site.")
|
||||
override suspend fun getListPage(page: Int, filter: MangaListFilter?): List<Manga> {
|
||||
val url = buildString {
|
||||
append("https://api.")
|
||||
append(domainNoSite)
|
||||
append("/query?adult=true&query_string=")
|
||||
when (filter) {
|
||||
is MangaListFilter.Search -> {
|
||||
append(filter.query.urlEncoded())
|
||||
}
|
||||
|
||||
is MangaListFilter.Advanced -> {
|
||||
|
||||
filter.states.oneOrThrowIfMany()?.let {
|
||||
append("&series_status=")
|
||||
append(
|
||||
when (it) {
|
||||
MangaState.ONGOING -> "Ongoing"
|
||||
MangaState.FINISHED -> "Completed"
|
||||
MangaState.ABANDONED -> "Dropped"
|
||||
MangaState.PAUSED -> "Hiatus"
|
||||
else -> ""
|
||||
},
|
||||
)
|
||||
|
||||
}
|
||||
append("&orderBy=")
|
||||
when (filter.sortOrder) {
|
||||
SortOrder.POPULARITY -> append("total_views&order=desc")
|
||||
SortOrder.UPDATED -> append("latest&order=desc")
|
||||
SortOrder.NEWEST -> append("created_at&order=desc")
|
||||
SortOrder.ALPHABETICAL -> append("title&order=asc")
|
||||
SortOrder.ALPHABETICAL_DESC -> append("title&order=desc")
|
||||
else -> append("latest&order=desc")
|
||||
}
|
||||
append("&series_type=All&perPage=")
|
||||
append(pageSize)
|
||||
append("&tags_ids=")
|
||||
append("[".urlEncoded())
|
||||
filter.tags.joinTo(this, ",") { it.key }
|
||||
append("]".urlEncoded())
|
||||
|
||||
}
|
||||
|
||||
null -> {}
|
||||
}
|
||||
append("&page=")
|
||||
append(page)
|
||||
}
|
||||
val json = webClient.httpGet(url).parseJson()
|
||||
return json.getJSONArray("data").mapJSON { j ->
|
||||
val slug = j.getString("series_slug")
|
||||
val urlManga = "https://$domain/$pathManga/$slug"
|
||||
val cover = if (j.getString("thumbnail").contains('/')) {
|
||||
j.getString("thumbnail")
|
||||
} else {
|
||||
"https://api.$domainNoSite/${j.getString("thumbnail")}"
|
||||
}
|
||||
Manga(
|
||||
id = j.getLong("id"),
|
||||
title = j.getString("title"),
|
||||
altTitle = null,
|
||||
url = urlManga.toRelativeUrl(domain),
|
||||
publicUrl = urlManga,
|
||||
rating = j.getFloatOrDefault("rating", RATING_UNKNOWN) / 5f,
|
||||
isNsfw = isNsfwSource,
|
||||
coverUrl = cover,
|
||||
tags = setOf(),
|
||||
state = when (j.getStringOrNull("status")) {
|
||||
"Ongoing" -> MangaState.ONGOING
|
||||
"Completed" -> MangaState.FINISHED
|
||||
"Dropped" -> MangaState.ABANDONED
|
||||
"Hiatus" -> MangaState.PAUSED
|
||||
else -> null
|
||||
},
|
||||
author = j.getStringOrNull("author"),
|
||||
source = source,
|
||||
description = j.getString("description"),
|
||||
)
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
override suspend fun getDetails(manga: Manga): Manga {
|
||||
val url = buildString {
|
||||
append("https://api.")
|
||||
append(domainNoSite)
|
||||
append("/chapter/query?perPage=9999&series_id=")
|
||||
append(manga.id)
|
||||
}
|
||||
val json = webClient.httpGet(url).parseJson()
|
||||
val dateFormat = SimpleDateFormat(datePattern, Locale.ENGLISH)
|
||||
|
||||
val chaptersJsonArray = json.getJSONArray("data")
|
||||
var totalChapters = json.getJSONObject("meta").getInt("total").toFloat()
|
||||
val chapters = chaptersJsonArray.mapJSON { j ->
|
||||
val slug = j.getJSONObject("series").getString("series_slug")
|
||||
val chapterUrl = "https://$domain/$pathManga/$slug/${j.getString("chapter_slug")}"
|
||||
val date = j.getString("created_at").substringBeforeLast("T")
|
||||
MangaChapter(
|
||||
id = j.getLong("id"),
|
||||
url = chapterUrl,
|
||||
name = j.getString("chapter_name"),
|
||||
number = totalChapters--,
|
||||
volume = 0,
|
||||
branch = null,
|
||||
uploadDate = dateFormat.tryParse(date),
|
||||
scanlator = null,
|
||||
source = source,
|
||||
)
|
||||
}
|
||||
|
||||
return manga.copy(
|
||||
chapters = chapters.reversed(),
|
||||
)
|
||||
}
|
||||
|
||||
override suspend fun getAvailableTags(): Set<MangaTag> {
|
||||
val doc = webClient.httpGet("https://$domain/comics").parseHtml()
|
||||
|
||||
val regex = Regex("\"tags\\\\.*?(\\[.+?])")
|
||||
val tags = doc.select("script").firstNotNullOf { script ->
|
||||
regex.find(script.html())?.groupValues?.getOrNull(1)
|
||||
}.unescapeJson()
|
||||
return JSONArray(tags).mapJSONToSet {
|
||||
MangaTag(
|
||||
key = it.getInt("id").toString(),
|
||||
title = it.getString("name").toTitleCase(sourceLocale),
|
||||
source = source,
|
||||
)
|
||||
}
|
||||
}
|
||||
internal class ModeScanlator(context: MangaLoaderContext) :
|
||||
HeanCms(context, MangaParserSource.MODESCANLATOR, "site.modescanlator.net") {
|
||||
override val apiPath = "api.modescanlator.net"
|
||||
}
|
||||
|
||||
|
||||
Loading…
Reference in New Issue