forked from Decentrala/chatbot
fix wiki commands
This commit is contained in:
commit
c011383f0e
@ -21,5 +21,6 @@ def command(msg, rcpt):
|
|||||||
response = client.chat(model='llama2-uncensored:latest', messages=[{'role':'user','content':f'{msg[4:]}'}])
|
response = client.chat(model='llama2-uncensored:latest', messages=[{'role':'user','content':f'{msg[4:]}'}])
|
||||||
return(response['message']['content'])
|
return(response['message']['content'])
|
||||||
elif msg.startswith("!wiki"):
|
elif msg.startswith("!wiki"):
|
||||||
return sf.query_external_website("https://en.wikipedia.org/wiki/", msg.split(" ")[1])
|
cmd, query = msg.split(" ", 1)
|
||||||
|
return sf.query_external_website("https://en.wikipedia.org", "/wiki/" + query)
|
||||||
|
|
||||||
|
@ -1,12 +1,19 @@
|
|||||||
import requests
|
import requests
|
||||||
from bs4 import BeautifulSoup
|
from bs4 import BeautifulSoup
|
||||||
|
from urllib.parse import quote
|
||||||
|
|
||||||
def query_external_website(base_url, query):
|
def query_external_website(base_url, query):
|
||||||
try:
|
try:
|
||||||
page = requests.get(base_url + query)
|
page = requests.get(base_url + quote(query))
|
||||||
soup = BeautifulSoup(page.content, "html.parser")
|
soup = BeautifulSoup(page.content, "html.parser")
|
||||||
title = soup.find("span", class_="mw-page-title-main").text
|
title = soup.find(id="firstHeading").text
|
||||||
content = soup.find(id="mw-content-text").select("p")[2].text
|
mainContentElement = soup.find(id="mw-content-text")
|
||||||
return "\nTITLE:\n" + title + "\n\nCONTENT:\n" + content + "\n\nFULL LINK:\n" + base_url + query
|
if "This page is a redirect" in mainContentElement.text:
|
||||||
except:
|
redirectLink = mainContentElement.find(class_="redirectMsg").find_all("a")[0]["href"]
|
||||||
return "Can't parse search result :("
|
return query_external_website(base_url, redirectLink)
|
||||||
|
content = next((paragraph for paragraph in mainContentElement.select("p") if not paragraph.has_attr("class")), None)
|
||||||
|
if content == None:
|
||||||
|
raise Exception("Can't parse search result :(")
|
||||||
|
return "\nTITLE:\n" + title + "\n\nCONTENT:\n" + content.text + "\n\nFULL LINK:\n" + base_url + quote(query)
|
||||||
|
except Exception as e:
|
||||||
|
return e
|
||||||
|
Loading…
Reference in New Issue
Block a user