from lxml import etree from nyaapy.magnet import magnet_builder from nyaapy.torrent import TorrentSite def nyaa_categories(b): c = b.replace("?c=", "") cats = c.split("_") cat = cats[0] sub_cat = cats[1] categories = { "1": { "name": "Anime", "sub_cats": { "1": "Anime Music Video", "2": "English-translated", "3": "Non-English-translated", "4": "Raw", }, }, "2": {"name": "Audio", "sub_cats": {"1": "Lossless", "2": "Lossy"}}, "3": { "name": "Literature", "sub_cats": { "1": "English-translated", "2": "Non-English-translated", "3": "Raw", }, }, "4": { "name": "Live Action", "sub_cats": { "1": "English-translated", "2": "Idol/Promotional Video", "3": "Non-English-translated", "4": "Raw", }, }, "5": {"name": "Pictures", "sub_cats": {"1": "Graphics", "2": "Photos"}}, "6": {"name": "Software", "sub_cats": {"1": "Applications", "2": "Games"}}, } try: category_name = ( f"{categories[cat]['name']} - {categories[cat]['sub_cats'][sub_cat]}" ) except KeyError: print("Unable to get Nyaa category name") return return category_name def parse_nyaa_rss(request_text, limit, site): """ Extracts torrent information from a given rss response. """ root = etree.fromstring(request_text) torrents = [] for item in root.xpath("channel/item")[:limit]: # Decide category. if site in [TorrentSite.NYAASI, TorrentSite.NYAALAND]: category = item.findtext("nyaa:categoryId", namespaces=item.nsmap) elif site in [TorrentSite.SUKEBEINYAASI, TorrentSite.SUKEBEINYAALAND]: category = item.findtext("nyaa:categoryId", namespaces=item.nsmap) else: raise ValueError("Unknown TorrentSite received!") try: is_remake = item.findtext("nyaa:remake", namespaces=item.nsmap) == "Yes" is_trusted = item.findtext("nyaa:trusted", namespaces=item.nsmap) == "Yes" item_type = ( "remake" if is_remake else "trusted" if is_trusted else "default" ) torrent = { "id": item.findtext("guid").split("/")[-1], "category": category, "url": item.findtext("guid"), "name": item.findtext("title"), "download_url": item.findtext("link"), "magnet": magnet_builder( item.findtext("nyaa:infoHash", namespaces=item.nsmap), item.findtext("title"), ), "size": item.findtext("nyaa:size", namespaces=item.nsmap), "date": item.findtext("pubDate"), "seeders": item.findtext("nyaa:seeders", namespaces=item.nsmap), "leechers": item.findtext("nyaa:leechers", namespaces=item.nsmap), "completed_downloads": None, "type": item_type, } torrents.append(torrent) except IndexError: pass return torrents def parse_nyaa(request_text, limit, site): parser = etree.HTMLParser() tree = etree.fromstring(request_text, parser) # Put proper domain here. uri = site.value torrents = [] # Going through table rows for tr in tree.xpath("//tbody//tr")[:limit]: block = [] for td in tr.xpath("./td"): for link in td.xpath("./a"): href = link.attrib.get("href").split("/")[-1] # Only caring about non-comment pages. if href[-9:] != "#comments": block.append(href) if link.text and link.text.strip(): block.append(link.text.strip()) if td.text is not None and td.text.strip(): block.append(td.text.strip()) # Add type of torrent based on tr class. if tr.attrib.get("class") is not None: if "danger" in tr.attrib.get("class"): block.append("remake") elif "success" in tr.attrib.get("class"): block.append("trusted") else: block.append("default") else: block.append("default") # Decide category. if site in [TorrentSite.NYAASI, TorrentSite.NYAALAND]: category = nyaa_categories(block[0]) elif site is TorrentSite.SUKEBEINYAASI: category = sukebei_categories(block[0]) else: raise ValueError("Unknown TorrentSite received!") # Create torrent object try: torrent = { "id": block[1], "category": category, "url": "{}/view/{}".format(uri, block[1]), "name": block[2], "download_url": "{}/download/{}".format(uri, block[3]), "magnet": block[4], "size": block[5], "date": block[6], "seeders": block[7], "leechers": block[8], "completed_downloads": block[9], "type": block[10], } torrents.append(torrent) except IndexError: pass return torrents def parse_single(request_text, site): parser = etree.HTMLParser() tree = etree.fromstring(request_text, parser) # Put proper domain here. uri = site.value torrent = {} data = [] torrent_files = [] # Find basic uploader info & torrent stats for row in tree.xpath("//div[@class='row']"): for div_text in row.xpath("./div[@class='col-md-5']//text()"): d = div_text.strip() if d: data.append(d) # Find files, we need only text of the li element(s). # Sorry about Pycodestyle aka PEP8 (E501) error for el in tree.xpath("//div[contains(@class, 'torrent-file-list')]//li/text()"): if el.rstrip(): torrent_files.append(el) torrent["title"] = tree.xpath("//h3[@class='panel-title']/text()")[0].strip() torrent["category"] = data[0] torrent["uploader"] = data[4] torrent["uploader_profile"] = "{}/user/{}".format(uri, data[4]) torrent["website"] = data[6] torrent["size"] = data[8] torrent["date"] = data[3] torrent["seeders"] = data[5] torrent["leechers"] = data[7] torrent["completed"] = data[9] torrent["hash"] = data[10] torrent["files"] = torrent_files torrent["description"] = "" for s in tree.xpath("//div[@id='torrent-description']"): torrent["description"] += s.text return torrent def sukebei_categories(b): c = b.replace("?c=", "") cats = c.split("_") cat = cats[0] subcat = cats[1] categories = { "1": { "name": "Art", "subcats": { "1": "Anime", "2": "Doujinshi", "3": "Games", "4": "Manga", "5": "Pictures", }, }, "2": { "name": "Real Life", "subcats": {"1": "Photobooks & Pictures", "2": "Videos"}, }, } try: category_name = ( f"{categories[cat]['name']} - {categories[cat]['subcats'][subcat]}" ) except KeyError: print("Unable to get Sukebei category name") return return category_name