rss-tools/bin/rssfind.py

89 lines
2.1 KiB
Python

#!/usr/bin/python3
import sys
import urllib.parse
from optparse import OptionParser
import feedparser
import orjson as json
import requests
from bs4 import BeautifulSoup as bs4
def findfeeds(url=None, disable_strict=False):
if url is None:
return None
raw = requests.get(url).text
results = []
discovered_feeds = []
html = bs4(raw, features="lxml")
feed_urls = html.findAll("link", rel="alternate")
if feed_urls:
for f in feed_urls:
tag = f.get("type", None)
if tag:
if "feed" in tag or "rss" in tag or "xml" in tag:
href = f.get("href", None)
if href:
discovered_feeds.append(href)
parsed_url = urllib.parse.urlparse(url)
base = f"{parsed_url.scheme}://{parsed_url.hostname}"
ahreftags = html.findAll("a")
for a in ahreftags:
href = a.get("href", None)
if href:
if "feed" in href or "rss" in href or "xml" in href:
discovered_feeds.append(f"{base}{href}")
for url in list(set(discovered_feeds)):
f = feedparser.parse(url)
if f.entries:
if url not in results:
results.append(url)
if disable_strict:
return list(set(discovered_feeds))
else:
return results
version = "0.2"
feedparser.USER_AGENT = (
"rssfind.py " + version + " +https://github.com/adulau/rss-tools"
)
usage = "Find RSS or Atom feeds from an URL\nusage: %prog [options]"
parser = OptionParser(usage)
parser.add_option(
"-l",
"--link",
dest="link",
help="http link where to find one or more feed source(s)",
)
parser.add_option(
"-d",
"--disable-strict",
action="store_false",
default=False,
help="Include empty feeds in the list, default strict is enabled",
)
(options, args) = parser.parse_args()
if not options.link:
print("URL missing")
parser.print_help()
sys.exit(0)
print(
json.dumps(findfeeds(options.link, disable_strict=options.disable_strict)).decode(
"utf-8"
)
)