mirror of
https://github.com/myned/modufur.git
synced 2024-12-25 06:37:29 +00:00
Refactor reverse commands, add Kheina database, and remove Harry.lu
This commit is contained in:
parent
c121371740
commit
ce132fbc64
2 changed files with 107 additions and 95 deletions
|
@ -106,23 +106,6 @@ class MsG(cmds.Cog):
|
||||||
|
|
||||||
return args
|
return args
|
||||||
|
|
||||||
def _get_icon(self, score):
|
|
||||||
if score is 'SauceNAO':
|
|
||||||
return 'https://d2.alternativeto.net/dist/icons/saucenao_23437.png?width=64&height=64&mode=crop&upscale=false'
|
|
||||||
elif score < 0:
|
|
||||||
return 'https://emojipedia-us.s3.amazonaws.com/thumbs/320/twitter/103/pouting-face_1f621.png'
|
|
||||||
elif score == 0:
|
|
||||||
return 'https://emojipedia-us.s3.amazonaws.com/thumbs/320/mozilla/36/pile-of-poo_1f4a9.png'
|
|
||||||
elif 10 > score > 0:
|
|
||||||
return 'https://emojipedia-us.s3.amazonaws.com/thumbs/320/twitter/103/white-medium-star_2b50.png'
|
|
||||||
elif 50 > score >= 10:
|
|
||||||
return 'https://emojipedia-us.s3.amazonaws.com/thumbs/320/twitter/103/glowing-star_1f31f.png'
|
|
||||||
elif 100 > score >= 50:
|
|
||||||
return 'https://emojipedia-us.s3.amazonaws.com/thumbs/320/twitter/103/dizzy-symbol_1f4ab.png'
|
|
||||||
elif score >= 100:
|
|
||||||
return 'https://emojipedia-us.s3.amazonaws.com/thumbs/320/twitter/103/sparkles_2728.png'
|
|
||||||
return None
|
|
||||||
|
|
||||||
async def _send_hearts(self):
|
async def _send_hearts(self):
|
||||||
while self.hearting:
|
while self.hearting:
|
||||||
temp = await self.heartqueue.get()
|
temp = await self.heartqueue.get()
|
||||||
|
@ -384,7 +367,7 @@ class MsG(cmds.Cog):
|
||||||
|
|
||||||
# Reverse image searches a linked image using the public iqdb
|
# Reverse image searches a linked image using the public iqdb
|
||||||
@cmds.cooldown(1, 5, cmds.BucketType.member)
|
@cmds.cooldown(1, 5, cmds.BucketType.member)
|
||||||
@cmds.command(name='reverse', aliases=['rev', 'ris'], brief='Reverse image search from e621', description='NSFW\nReverse-search an image with given URL')
|
@cmds.command(name='reverse', aliases=['rev', 'ris'], brief='Reverse image search from Kheina and SauceNAO', description='NSFW\nReverse-search an image with given URL')
|
||||||
async def reverse(self, ctx, *args):
|
async def reverse(self, ctx, *args):
|
||||||
try:
|
try:
|
||||||
kwargs = u.get_kwargs(ctx, args)
|
kwargs = u.get_kwargs(ctx, args)
|
||||||
|
@ -400,19 +383,15 @@ class MsG(cmds.Cog):
|
||||||
async with ctx.channel.typing():
|
async with ctx.channel.typing():
|
||||||
for url in urls:
|
for url in urls:
|
||||||
try:
|
try:
|
||||||
post, source, similarity = await scraper.get_post(url)
|
result = await scraper.get_post(url)
|
||||||
|
|
||||||
embed = d.Embed(
|
embed = d.Embed(
|
||||||
title=', '.join(post['artist']),
|
title=result['artist'],
|
||||||
url=source,
|
url=result['source'],
|
||||||
color=ctx.me.color if isinstance(ctx.channel, d.TextChannel) else u.color)
|
color=ctx.me.color if isinstance(ctx.channel, d.TextChannel) else u.color)
|
||||||
embed.set_image(url=post['file_url'])
|
embed.set_image(url=result['thumbnail'])
|
||||||
embed.set_author(
|
embed.set_author(name=result['similarity'] + '% Match', icon_url=ctx.author.avatar_url)
|
||||||
name=similarity,
|
embed.set_footer(text=result['database'])
|
||||||
icon_url=ctx.author.avatar_url)
|
|
||||||
embed.set_footer(
|
|
||||||
text=post['score'],
|
|
||||||
icon_url=self._get_icon(post['score']))
|
|
||||||
|
|
||||||
await ctx.send(embed=embed)
|
await ctx.send(embed=embed)
|
||||||
|
|
||||||
|
@ -474,19 +453,15 @@ class MsG(cmds.Cog):
|
||||||
for message, urls in links.items():
|
for message, urls in links.items():
|
||||||
for url in urls:
|
for url in urls:
|
||||||
try:
|
try:
|
||||||
post, source, similarity = await scraper.get_post(url)
|
result = await scraper.get_post(url)
|
||||||
|
|
||||||
embed = d.Embed(
|
embed = d.Embed(
|
||||||
title=', '.join(post['artist']),
|
title=result['artist'],
|
||||||
url=source,
|
url=result['source'],
|
||||||
color=ctx.me.color if isinstance(ctx.channel, d.TextChannel) else u.color)
|
color=ctx.me.color if isinstance(ctx.channel, d.TextChannel) else u.color)
|
||||||
embed.set_image(url=post['file_url'])
|
embed.set_image(url=result['thumbnail'])
|
||||||
embed.set_author(
|
embed.set_author(name=result['similarity'] + '% Match', icon_url=ctx.author.avatar_url)
|
||||||
name=similarity,
|
embed.set_footer(text=result['database'])
|
||||||
icon_url=message.author.avatar_url)
|
|
||||||
embed.set_footer(
|
|
||||||
text=post['score'],
|
|
||||||
icon_url=self._get_icon(post['score']))
|
|
||||||
|
|
||||||
await dest.send(embed=embed)
|
await dest.send(embed=embed)
|
||||||
await message.add_reaction('\N{WHITE HEAVY CHECK MARK}')
|
await message.add_reaction('\N{WHITE HEAVY CHECK MARK}')
|
||||||
|
@ -536,19 +511,15 @@ class MsG(cmds.Cog):
|
||||||
async with message.channel.typing():
|
async with message.channel.typing():
|
||||||
for url in urls:
|
for url in urls:
|
||||||
try:
|
try:
|
||||||
post, source, similarity = await scraper.get_post(url)
|
result = await scraper.get_post(url)
|
||||||
|
|
||||||
embed = d.Embed(
|
embed = d.Embed(
|
||||||
title=', '.join(post['artist']),
|
title=result['artist'],
|
||||||
url=source,
|
url=result['source'],
|
||||||
color=message.channel.guild.me.color if isinstance(message.channel, d.TextChannel) else u.color)
|
color=message.me.color if isinstance(message.channel, d.TextChannel) else u.color)
|
||||||
embed.set_image(url=post['file_url'])
|
embed.set_image(url=result['thumbnail'])
|
||||||
embed.set_author(
|
embed.set_author(name=result['similarity'] + '% Match', icon_url=message.author.avatar_url)
|
||||||
name=similarity,
|
embed.set_footer(text=result['database'])
|
||||||
icon_url=message.author.avatar_url)
|
|
||||||
embed.set_footer(
|
|
||||||
text=post['score'],
|
|
||||||
icon_url=self._get_icon(post['score']))
|
|
||||||
|
|
||||||
await message.channel.send(embed=embed)
|
await message.channel.send(embed=embed)
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
import asyncio
|
import aiohttp
|
||||||
|
import ast
|
||||||
import re
|
import re
|
||||||
|
|
||||||
from bs4 import BeautifulSoup
|
from bs4 import BeautifulSoup
|
||||||
|
@ -9,58 +10,98 @@ from misc import exceptions as exc
|
||||||
from utils import utils as u
|
from utils import utils as u
|
||||||
|
|
||||||
|
|
||||||
async def get_post(url):
|
# async def get_harry(url):
|
||||||
try:
|
# content = await u.fetch('https://iqdb.harry.lu', params={'url': url})
|
||||||
image = await u.fetch(url, response=True)
|
# soup = BeautifulSoup(content, 'html.parser')
|
||||||
filesize = int(image.headers['Content-Length'])
|
#
|
||||||
if filesize > 8192 * 1024:
|
# if soup.find('div', id='show1').string is 'Not the right one? ':
|
||||||
raise exc.SizeError(size(filesize, system=alternative))
|
# parent = soup.find('th', string='Probable match:').parent.parent
|
||||||
|
#
|
||||||
|
# post = await u.fetch(
|
||||||
|
# 'https://e621.net/post/show.json',
|
||||||
|
# params={'id': re.search('show/([0-9]+)', parent.tr.td.a.get('href')).group(1)},
|
||||||
|
# json=True)
|
||||||
|
# if (post['status'] == 'deleted'):
|
||||||
|
# post = await u.fetch(
|
||||||
|
# 'https://e621.net/post/show.json',
|
||||||
|
# params={'id': re.search('#(\\d+)', post['delreason']).group(1)},
|
||||||
|
# json=True)
|
||||||
|
#
|
||||||
|
# result = {
|
||||||
|
# 'source': f'https://e621.net/post/show/{post["id"]}',
|
||||||
|
# 'artist': ', '.join(post['artist']),
|
||||||
|
# 'thumbnail': parent.td.a.img.get('src'),
|
||||||
|
# 'similarity': re.search('\\d+', parent.tr[4].td.string).group(0),
|
||||||
|
# 'database': 'Harry.lu'
|
||||||
|
# }
|
||||||
|
#
|
||||||
|
# return result
|
||||||
|
# else:
|
||||||
|
# return False
|
||||||
|
|
||||||
content = await u.fetch('http://iqdb.harry.lu', params={'url': url})
|
|
||||||
|
async def get_kheina(url):
|
||||||
|
content = await u.fetch('https://kheina.com', params={'url': url})
|
||||||
soup = BeautifulSoup(content, 'html.parser')
|
soup = BeautifulSoup(content, 'html.parser')
|
||||||
source = soup.find_all('a', limit=2)[1].get('href')
|
|
||||||
|
|
||||||
if source != '#':
|
results = ast.literal_eval(soup.find('data', id='results').string)[-1]
|
||||||
ident = re.search('show/([0-9]+)', source).group(1)
|
iqdbdata = ast.literal_eval(soup.find('data', id='iqdbdata').string)[0]
|
||||||
post = await u.fetch('http://e621.net/post/show.json', params={'id': ident}, json=True)
|
|
||||||
if (post['status'] == 'deleted'):
|
|
||||||
ident = re.search('#(\\d+)', post['delreason']).group(1)
|
|
||||||
post = await u.fetch('http://e621.net/post/show.json', params={'id': ident}, json=True)
|
|
||||||
source = f'https://e621.net/post/show/{post["id"]}'
|
|
||||||
similarity = re.search('\\d+', soup.find(string=re.compile('similarity'))).group(0)
|
|
||||||
|
|
||||||
return post, source, similarity + '% Match'
|
result = {
|
||||||
else:
|
'source': results[3],
|
||||||
raise IndexError
|
'artist': results[4],
|
||||||
|
'thumbnail': f'https://f002.backblazeb2.com/file/kheinacom/{results[1]}.jpg',
|
||||||
|
'similarity': str(int(float(iqdbdata['similarity']))),
|
||||||
|
'database': 'Kheina'
|
||||||
|
}
|
||||||
|
|
||||||
except IndexError:
|
return result
|
||||||
|
|
||||||
|
|
||||||
|
async def get_saucenao(url):
|
||||||
content = await u.fetch(
|
content = await u.fetch(
|
||||||
'https://saucenao.com/search.php',
|
'https://saucenao.com/search.php',
|
||||||
params={
|
params={'url': url, 'api_key': u.config['saucenao_api'], 'output_type': 2},
|
||||||
'url': url,
|
|
||||||
'api_key': u.config['saucenao_api'],
|
|
||||||
'output_type': 2},
|
|
||||||
json=True)
|
json=True)
|
||||||
result = content['results'][0]
|
results = content['results'][0]
|
||||||
if 'author_name' in result['data']:
|
for i in range(len(content['results'])):
|
||||||
|
if 'e621' in content['results'][i]['header']['index_name']:
|
||||||
|
results = content['results'][i]
|
||||||
|
|
||||||
|
if 'author_name' in results['data']:
|
||||||
artist = 'author_name'
|
artist = 'author_name'
|
||||||
elif 'member_name' in result['data']:
|
elif 'member_name' in results['data']:
|
||||||
artist = 'member_name'
|
artist = 'member_name'
|
||||||
else:
|
else:
|
||||||
artist = 'creator'
|
artist = 'creator'
|
||||||
post = {
|
|
||||||
'file_url': result['header']['thumbnail'],
|
|
||||||
'artist': [result['data'][artist]],
|
|
||||||
'score': 'SauceNAO'}
|
|
||||||
source = result['data']['ext_urls'][0]
|
|
||||||
similarity = re.search('(\\d+)\\.', result['header']['similarity']).group(1)
|
|
||||||
|
|
||||||
if int(similarity) >= 55:
|
result = {
|
||||||
return post, source, similarity + '% Match'
|
'source': results['data']['ext_urls'][0],
|
||||||
|
'artist': results['data'][artist],
|
||||||
|
'thumbnail': results['header']['thumbnail'],
|
||||||
|
'similarity': str(int(float(results['header']['similarity']))),
|
||||||
|
'database': 'SauceNAO'
|
||||||
|
}
|
||||||
|
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
|
async def get_post(url):
|
||||||
|
try:
|
||||||
|
content = await u.fetch(url, response=True)
|
||||||
|
filesize = int(content.headers['Content-Length'])
|
||||||
|
if filesize > 8192 * 1024:
|
||||||
|
raise exc.SizeError(size(filesize, system=alternative))
|
||||||
|
|
||||||
|
result = await get_kheina(url)
|
||||||
|
if int(result['similarity']) < 55:
|
||||||
|
result = await get_saucenao(url)
|
||||||
|
if int(result['similarity']) < 55:
|
||||||
raise exc.MatchError(re.search('\\/([^\\/]+)$', url).group(1))
|
raise exc.MatchError(re.search('\\/([^\\/]+)$', url).group(1))
|
||||||
|
|
||||||
except (AttributeError, ValueError, KeyError):
|
return result
|
||||||
|
|
||||||
|
except aiohttp.InvalidURL:
|
||||||
raise exc.MissingArgument
|
raise exc.MissingArgument
|
||||||
|
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue