mirror of
https://github.com/myned/modufur.git
synced 2024-11-01 21:02:38 +00:00
Merge branch 'dev'
This commit is contained in:
commit
25c9bbafab
5 changed files with 336 additions and 110 deletions
|
@ -1,5 +1,6 @@
|
||||||
import asyncio
|
import asyncio
|
||||||
import json
|
import json
|
||||||
|
import re
|
||||||
import traceback as tb
|
import traceback as tb
|
||||||
from contextlib import suppress
|
from contextlib import suppress
|
||||||
|
|
||||||
|
@ -70,6 +71,25 @@ class MsG:
|
||||||
except exc.MissingArgument:
|
except exc.MissingArgument:
|
||||||
await ctx.send('❌ **No tags given.**', delete_after=10)
|
await ctx.send('❌ **No tags given.**', delete_after=10)
|
||||||
|
|
||||||
|
@commands.command(name='getimage', aliases=['geti', 'gi'])
|
||||||
|
@checks.del_ctx()
|
||||||
|
async def get_image(self, ctx, *urls):
|
||||||
|
try:
|
||||||
|
if not urls:
|
||||||
|
raise exc.MissingArgument
|
||||||
|
|
||||||
|
for url in urls:
|
||||||
|
try:
|
||||||
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
|
await ctx.send('✅ {}'.format(await scraper.get_image(url)))
|
||||||
|
|
||||||
|
finally:
|
||||||
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
|
||||||
|
except exc.MissingArgument:
|
||||||
|
await ctx.send('❌ **Invalid url or file.**', delete_after=10)
|
||||||
|
|
||||||
# Reverse image searches a linked image using the public iqdb
|
# Reverse image searches a linked image using the public iqdb
|
||||||
@commands.command(name='reverse', aliases=['rev', 'ris'], brief='e621 Reverse image search', description='e621 | NSFW\nReverse-search an image with given URL')
|
@commands.command(name='reverse', aliases=['rev', 'ris'], brief='e621 Reverse image search', description='e621 | NSFW\nReverse-search an image with given URL')
|
||||||
@checks.del_ctx()
|
@checks.del_ctx()
|
||||||
|
@ -78,27 +98,26 @@ class MsG:
|
||||||
if not urls and not ctx.message.attachments:
|
if not urls and not ctx.message.attachments:
|
||||||
raise exc.MissingArgument
|
raise exc.MissingArgument
|
||||||
|
|
||||||
await ctx.trigger_typing()
|
|
||||||
|
|
||||||
for url in urls:
|
for url in urls:
|
||||||
try:
|
try:
|
||||||
await ctx.trigger_typing()
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
await ctx.send('✅ **Probable match:**\n{}'.format(await scraper.check_match(url)))
|
await ctx.send('✅ **Probable match:**\n{}'.format(await scraper.get_post(url)))
|
||||||
|
|
||||||
except exc.MatchError as e:
|
except exc.MatchError as e:
|
||||||
await ctx.send('❌ **No probable match for** `{}`**.** Make sure URLs direct to an image file.'.format(e), delete_after=10)
|
await ctx.send('❌ **No probable match for:** `{}`'.format(e), delete_after=10)
|
||||||
|
|
||||||
finally:
|
finally:
|
||||||
await asyncio.sleep(self.RATE_LIMIT)
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
|
||||||
for attachment in ctx.message.attachments:
|
for attachment in ctx.message.attachments:
|
||||||
try:
|
try:
|
||||||
await ctx.trigger_typing()
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
await ctx.send('✅ **Probable match:**\n{}'.format(await scraper.check_match(attachment.url)))
|
await ctx.send('✅ **Probable match:**\n{}'.format(await scraper.get_post(attachment.url)))
|
||||||
|
|
||||||
except exc.MatchError as e:
|
except exc.MatchError as e:
|
||||||
await ctx.send('❌ **No probable match for** `{}`**.** Make sure URLs direct to an image file.'.format(e), delete_after=10)
|
await ctx.send('❌ **No probable match for:** `{}`'.format(e), delete_after=10)
|
||||||
|
|
||||||
finally:
|
finally:
|
||||||
await asyncio.sleep(self.RATE_LIMIT)
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
@ -106,6 +125,181 @@ class MsG:
|
||||||
except exc.MissingArgument:
|
except exc.MissingArgument:
|
||||||
await ctx.send('❌ **Invalid url or file.**', delete_after=10)
|
await ctx.send('❌ **Invalid url or file.**', delete_after=10)
|
||||||
|
|
||||||
|
@commands.command(name='reverseall', aliases=['revall', 'risall', 'rall'])
|
||||||
|
@checks.del_ctx()
|
||||||
|
@commands.has_permissions(manage_channels=True)
|
||||||
|
async def reverse_image_search_all(self, ctx, arg=None, limit=1):
|
||||||
|
urls = []
|
||||||
|
attachments = []
|
||||||
|
delete = False
|
||||||
|
|
||||||
|
try:
|
||||||
|
if arg == '-d' or arg == '-del' or arg == '-delete':
|
||||||
|
delete = True
|
||||||
|
elif arg is not None:
|
||||||
|
limit = int(arg)
|
||||||
|
|
||||||
|
async for message in ctx.channel.history(limit=limit + 1):
|
||||||
|
if re.search('(http[a-z]?:\/\/[^ ]*\.(?:gif|png|jpg|jpeg))', message.content) is not None:
|
||||||
|
urls.append(message)
|
||||||
|
elif message.attachments:
|
||||||
|
attachments.append(message)
|
||||||
|
|
||||||
|
if not urls and not attachments:
|
||||||
|
raise exc.NotFound
|
||||||
|
|
||||||
|
for message in urls:
|
||||||
|
for match in re.finditer('(http[a-z]?:\/\/[^ ]*\.(?:gif|png|jpg|jpeg))', message.content):
|
||||||
|
try:
|
||||||
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
|
await ctx.send('✅ **Probable match from {}:**\n{}'.format(message.author.display_name, await scraper.get_post(match.group(0))))
|
||||||
|
|
||||||
|
if delete:
|
||||||
|
with suppress(err.NotFound):
|
||||||
|
await message.delete()
|
||||||
|
|
||||||
|
except exc.MatchError as e:
|
||||||
|
await ctx.send('❌ **No probable match for:** `{}`'.format(e), delete_after=10)
|
||||||
|
|
||||||
|
finally:
|
||||||
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
|
||||||
|
for message in attachments:
|
||||||
|
for attachment in message.attachments:
|
||||||
|
try:
|
||||||
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
|
await ctx.send('✅ **Probable match from {}:**\n{}'.format(message.author.display_name, await scraper.get_post(attachment.url)))
|
||||||
|
|
||||||
|
if delete:
|
||||||
|
with suppress(err.NotFound):
|
||||||
|
await message.delete()
|
||||||
|
|
||||||
|
except exc.MatchError as e:
|
||||||
|
await ctx.send('❌ **No probable match for:** `{}`'.format(e), delete_after=10)
|
||||||
|
|
||||||
|
finally:
|
||||||
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
|
||||||
|
except exc.NotFound:
|
||||||
|
await ctx.send('❌ **No matches found.**', delete_after=10)
|
||||||
|
except ValueError:
|
||||||
|
await ctx.send('❌ **Invalid limit.**', delete_after=10)
|
||||||
|
|
||||||
|
@commands.command(name='qualitysearch', aliases=['qrev', 'qis'])
|
||||||
|
@checks.del_ctx()
|
||||||
|
async def quality_reverse_image_search(self, ctx, *urls):
|
||||||
|
try:
|
||||||
|
if not urls and not ctx.message.attachments:
|
||||||
|
raise exc.MissingArgument
|
||||||
|
|
||||||
|
for url in urls:
|
||||||
|
try:
|
||||||
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
|
post = await scraper.get_post(url)
|
||||||
|
|
||||||
|
await ctx.send('✅ **Probable match:**\n{}'.format(await scraper.get_image(post)))
|
||||||
|
|
||||||
|
except exc.MatchError as e:
|
||||||
|
await ctx.send('❌ **No probable match for:** `{}`'.format(e), delete_after=10)
|
||||||
|
|
||||||
|
finally:
|
||||||
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
|
||||||
|
for attachment in ctx.message.attachments:
|
||||||
|
try:
|
||||||
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
|
post = await scraper.get_post(attachment.url)
|
||||||
|
|
||||||
|
await ctx.send('✅ **Probable match:**\n{}'.format(await scraper.get_image(post)))
|
||||||
|
|
||||||
|
except exc.MatchError as e:
|
||||||
|
await ctx.send('❌ **No probable match for:** `{}`'.format(e), delete_after=10)
|
||||||
|
|
||||||
|
finally:
|
||||||
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
|
||||||
|
except exc.MissingArgument:
|
||||||
|
await ctx.send('❌ **Invalid url or file.**', delete_after=10)
|
||||||
|
|
||||||
|
@commands.command(name='qualityall', aliases=['qrevall', 'qisall', 'qall'])
|
||||||
|
@checks.del_ctx()
|
||||||
|
@commands.has_permissions(manage_channels=True)
|
||||||
|
async def quality_reverse_image_search_all(self, ctx, arg=None, limit=1):
|
||||||
|
urls = []
|
||||||
|
attachments = []
|
||||||
|
delete = False
|
||||||
|
|
||||||
|
try:
|
||||||
|
if arg == '-d' or arg == '-del' or arg == '-delete':
|
||||||
|
delete = True
|
||||||
|
elif arg is not None:
|
||||||
|
limit = int(arg)
|
||||||
|
|
||||||
|
async for message in ctx.channel.history(limit=limit + 1):
|
||||||
|
if re.search('(http[a-z]?:\/\/[^ ]*\.(?:gif|png|jpg|jpeg))', message.content) is not None:
|
||||||
|
urls.append(message)
|
||||||
|
elif message.attachments:
|
||||||
|
attachments.append(message)
|
||||||
|
|
||||||
|
if not urls and not attachments:
|
||||||
|
raise exc.NotFound
|
||||||
|
|
||||||
|
for message in urls:
|
||||||
|
for match in re.finditer('(http[a-z]?:\/\/[^ ]*\.(?:gif|png|jpg|jpeg))', message.content):
|
||||||
|
try:
|
||||||
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
|
post = await scraper.get_post(match.group(0))
|
||||||
|
|
||||||
|
await ctx.send('✅ **Probable match from {}:**\n{}'.format(message.author.display_name, await scraper.get_image(post)))
|
||||||
|
|
||||||
|
if delete:
|
||||||
|
with suppress(err.NotFound):
|
||||||
|
await message.delete()
|
||||||
|
|
||||||
|
except exc.MatchError as e:
|
||||||
|
await ctx.send('❌ **No probable match for:** `{}`'.format(e), delete_after=10)
|
||||||
|
|
||||||
|
finally:
|
||||||
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
|
||||||
|
for message in attachments:
|
||||||
|
for attachment in message.attachments:
|
||||||
|
try:
|
||||||
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
|
post = await scraper.get_post(attachment.url)
|
||||||
|
|
||||||
|
await ctx.send('✅ **Probable match from {}:**\n{}'.format(message.author.display_name, await scraper.get_image(post)))
|
||||||
|
|
||||||
|
if delete:
|
||||||
|
with suppress(err.NotFound):
|
||||||
|
await message.delete()
|
||||||
|
|
||||||
|
except exc.MatchError as e:
|
||||||
|
await ctx.send('❌ **No probable match for:** `{}`'.format(e), delete_after=10)
|
||||||
|
|
||||||
|
finally:
|
||||||
|
await asyncio.sleep(self.RATE_LIMIT)
|
||||||
|
|
||||||
|
except exc.NotFound:
|
||||||
|
await ctx.send('❌ **No matches found.**', delete_after=10)
|
||||||
|
except ValueError:
|
||||||
|
await ctx.send('❌ **Invalid limit.**', delete_after=10)
|
||||||
|
|
||||||
|
def get_favorites(self, ctx, args):
|
||||||
|
if '-f' in args or '-favs' in args or '-faves' in args or '-favorites' in args:
|
||||||
|
if self.favorites.get(ctx.author.id, {}).get('tags', set()):
|
||||||
|
args = ['~{}'.format(tag) for tag in self.favorites[ctx.author.id]['tags']]
|
||||||
|
else:
|
||||||
|
raise exc.FavoritesNotFound
|
||||||
|
|
||||||
|
return args
|
||||||
|
|
||||||
async def return_pool(self, *, ctx, booru='e621', query=[]):
|
async def return_pool(self, *, ctx, booru='e621', query=[]):
|
||||||
def on_message(msg):
|
def on_message(msg):
|
||||||
if msg.content.lower() == 'cancel' and msg.author is ctx.author and msg.channel is ctx.channel:
|
if msg.content.lower() == 'cancel' and msg.author is ctx.author and msg.channel is ctx.channel:
|
||||||
|
@ -153,7 +347,7 @@ class MsG:
|
||||||
@commands.command(name='pool', aliases=['e6pp'], brief='e621 pool paginator', description='e621 | NSFW\nShow pools in a page format', hidden=True)
|
@commands.command(name='pool', aliases=['e6pp'], brief='e621 pool paginator', description='e621 | NSFW\nShow pools in a page format', hidden=True)
|
||||||
@checks.del_ctx()
|
@checks.del_ctx()
|
||||||
async def pool_paginator(self, ctx, *kwords):
|
async def pool_paginator(self, ctx, *kwords):
|
||||||
def on_react(reaction, user):
|
def on_reaction(reaction, user):
|
||||||
if reaction.emoji == '🚫' and reaction.message.content == paginator.content and (user is ctx.author or user.id == u.config['owner_id']):
|
if reaction.emoji == '🚫' and reaction.message.content == paginator.content and (user is ctx.author or user.id == u.config['owner_id']):
|
||||||
raise exc.Abort
|
raise exc.Abort
|
||||||
elif reaction.emoji == '📁' and reaction.message.content == paginator.content and (user is ctx.author or user.id == u.config['owner_id']):
|
elif reaction.emoji == '📁' and reaction.message.content == paginator.content and (user is ctx.author or user.id == u.config['owner_id']):
|
||||||
|
@ -178,6 +372,7 @@ class MsG:
|
||||||
|
|
||||||
try:
|
try:
|
||||||
await ctx.trigger_typing()
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
pool, posts = await self.return_pool(ctx=ctx, booru='e621', query=kwords)
|
pool, posts = await self.return_pool(ctx=ctx, booru='e621', query=kwords)
|
||||||
keys = list(posts.keys())
|
keys = list(posts.keys())
|
||||||
values = list(posts.values())
|
values = list(posts.values())
|
||||||
|
@ -200,7 +395,7 @@ class MsG:
|
||||||
|
|
||||||
while not self.bot.is_closed():
|
while not self.bot.is_closed():
|
||||||
try:
|
try:
|
||||||
await self.bot.wait_for('reaction_add', check=on_react, timeout=10 * 60)
|
await self.bot.wait_for('reaction_add', check=on_reaction, timeout=10 * 60)
|
||||||
|
|
||||||
except exc.Left:
|
except exc.Left:
|
||||||
if c > 1:
|
if c > 1:
|
||||||
|
@ -326,7 +521,7 @@ class MsG:
|
||||||
@checks.del_ctx()
|
@checks.del_ctx()
|
||||||
@checks.is_nsfw()
|
@checks.is_nsfw()
|
||||||
async def e621_paginator(self, ctx, *args):
|
async def e621_paginator(self, ctx, *args):
|
||||||
def on_react(reaction, user):
|
def on_reaction(reaction, user):
|
||||||
if reaction.emoji == '🚫' and reaction.message.content == paginator.content and (user is ctx.author or user.id == u.config['owner_id']):
|
if reaction.emoji == '🚫' and reaction.message.content == paginator.content and (user is ctx.author or user.id == u.config['owner_id']):
|
||||||
raise exc.Abort
|
raise exc.Abort
|
||||||
elif reaction.emoji == '📁' and reaction.message.content == paginator.content and (user is ctx.author or user.id == u.config['owner_id']):
|
elif reaction.emoji == '📁' and reaction.message.content == paginator.content and (user is ctx.author or user.id == u.config['owner_id']):
|
||||||
|
@ -352,7 +547,10 @@ class MsG:
|
||||||
c = 1
|
c = 1
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
args = self.get_favorites(ctx, args)
|
||||||
|
|
||||||
await ctx.trigger_typing()
|
await ctx.trigger_typing()
|
||||||
|
|
||||||
posts = await self.check_return_posts(ctx=ctx, booru='e621', tags=args, limit=limit)
|
posts = await self.check_return_posts(ctx=ctx, booru='e621', tags=args, limit=limit)
|
||||||
keys = list(posts.keys())
|
keys = list(posts.keys())
|
||||||
values = list(posts.values())
|
values = list(posts.values())
|
||||||
|
@ -375,7 +573,7 @@ class MsG:
|
||||||
|
|
||||||
while not self.bot.is_closed():
|
while not self.bot.is_closed():
|
||||||
try:
|
try:
|
||||||
await self.bot.wait_for('reaction_add', check=on_react, timeout=10 * 60)
|
await self.bot.wait_for('reaction_add', check=on_reaction, timeout=10 * 60)
|
||||||
|
|
||||||
except exc.Left:
|
except exc.Left:
|
||||||
if c > 1:
|
if c > 1:
|
||||||
|
@ -431,19 +629,16 @@ class MsG:
|
||||||
|
|
||||||
except exc.Abort:
|
except exc.Abort:
|
||||||
await paginator.edit(content='🚫 **Exited paginator.**')
|
await paginator.edit(content='🚫 **Exited paginator.**')
|
||||||
|
|
||||||
except asyncio.TimeoutError:
|
except asyncio.TimeoutError:
|
||||||
await paginator.edit(content='❌ **Paginator timed out.**')
|
await paginator.edit(content='❌ **Paginator timed out.**')
|
||||||
|
|
||||||
except exc.NotFound as e:
|
except exc.NotFound as e:
|
||||||
await ctx.send('❌ `{}` **not found.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **not found.**'.format(e), delete_after=10)
|
||||||
|
|
||||||
except exc.TagBlacklisted as e:
|
except exc.TagBlacklisted as e:
|
||||||
await ctx.send('❌ `{}` **blacklisted.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **blacklisted.**'.format(e), delete_after=10)
|
||||||
|
|
||||||
except exc.TagBoundsError as e:
|
except exc.TagBoundsError as e:
|
||||||
await ctx.send('❌ `{}` **out of bounds.** Tags limited to 5, currently.'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **out of bounds.** Tags limited to 5, currently.'.format(e), delete_after=10)
|
||||||
|
except exc.FavoritesNotFound:
|
||||||
|
await ctx.send('❌ **You have no favorite tags.**', delete_after=10)
|
||||||
except exc.Timeout:
|
except exc.Timeout:
|
||||||
await ctx.send('❌ **Request timed out.**')
|
await ctx.send('❌ **Request timed out.**')
|
||||||
|
|
||||||
|
@ -458,26 +653,36 @@ class MsG:
|
||||||
if isinstance(error, errext.CheckFailure):
|
if isinstance(error, errext.CheckFailure):
|
||||||
return await ctx.send('⛔️ {} **is not an NSFW channel.**'.format(ctx.channel.mention), delete_after=10)
|
return await ctx.send('⛔️ {} **is not an NSFW channel.**'.format(ctx.channel.mention), delete_after=10)
|
||||||
|
|
||||||
|
def get_limit(self, args):
|
||||||
|
limit = 1
|
||||||
|
|
||||||
|
for arg in args:
|
||||||
|
if len(arg) == 1:
|
||||||
|
with suppress(ValueError):
|
||||||
|
if int(arg) <= 3 and int(arg) >= 1:
|
||||||
|
limit = int(arg)
|
||||||
|
args.remove(arg)
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
raise exc.BoundsError(arg)
|
||||||
|
|
||||||
|
return limit
|
||||||
|
|
||||||
# Searches for and returns images from e621.net given tags when not blacklisted
|
# Searches for and returns images from e621.net given tags when not blacklisted
|
||||||
@commands.command(aliases=['e6', '6'], brief='e621 | NSFW', description='e621 | NSFW\nTag-based search for e621.net\n\nYou can only search 5 tags and 6 images at once for now.\ne6 [tags...] ([# of images])')
|
@commands.group(aliases=['e6', '6'], brief='e621 | NSFW', description='e621 | NSFW\nTag-based search for e621.net\n\nYou can only search 5 tags and 6 images at once for now.\ne6 [tags...] ([# of images])')
|
||||||
@checks.del_ctx()
|
@checks.del_ctx()
|
||||||
@checks.is_nsfw()
|
@checks.is_nsfw()
|
||||||
async def e621(self, ctx, *args):
|
async def e621(self, ctx, *args):
|
||||||
args = list(args)
|
args = list(args)
|
||||||
limit = 1
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
args = self.get_favorites(ctx, args)
|
||||||
|
limit = self.get_limit(args)
|
||||||
|
|
||||||
await ctx.trigger_typing()
|
await ctx.trigger_typing()
|
||||||
# Checks for, defines, and removes limit from args
|
|
||||||
for arg in args:
|
|
||||||
if len(arg) == 1:
|
|
||||||
with suppress(ValueError):
|
|
||||||
if int(arg) <= 6 and int(arg) >= 1:
|
|
||||||
limit = int(arg)
|
|
||||||
args.remove(arg)
|
|
||||||
else:
|
|
||||||
raise exc.BoundsError(arg)
|
|
||||||
posts = await self.check_return_posts(ctx=ctx, booru='e621', tags=args, limit=limit)
|
posts = await self.check_return_posts(ctx=ctx, booru='e621', tags=args, limit=limit)
|
||||||
|
|
||||||
for ident, post in posts.items():
|
for ident, post in posts.items():
|
||||||
embed = d.Embed(title=post['author'], url='https://e621.net/post/show/{}'.format(ident),
|
embed = d.Embed(title=post['author'], url='https://e621.net/post/show/{}'.format(ident),
|
||||||
color=ctx.me.color).set_image(url=post['url'])
|
color=ctx.me.color).set_image(url=post['url'])
|
||||||
|
@ -485,20 +690,19 @@ class MsG:
|
||||||
url='https://e621.net/post?tags={}'.format(','.join(args)), icon_url=ctx.author.avatar_url)
|
url='https://e621.net/post?tags={}'.format(','.join(args)), icon_url=ctx.author.avatar_url)
|
||||||
embed.set_footer(
|
embed.set_footer(
|
||||||
text=str(ident), icon_url='http://lh6.ggpht.com/d3pNZNFCcJM8snBsRSdKUhR9AVBnJMcYYrR92RRDBOzCrxZMhuTeoGOQSmSEn7DAPQ=w300')
|
text=str(ident), icon_url='http://lh6.ggpht.com/d3pNZNFCcJM8snBsRSdKUhR9AVBnJMcYYrR92RRDBOzCrxZMhuTeoGOQSmSEn7DAPQ=w300')
|
||||||
|
|
||||||
await ctx.send(embed=embed)
|
await ctx.send(embed=embed)
|
||||||
|
|
||||||
except exc.TagBlacklisted as e:
|
except exc.TagBlacklisted as e:
|
||||||
await ctx.send('❌ `{}` **blacklisted.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **blacklisted.**'.format(e), delete_after=10)
|
||||||
|
|
||||||
except exc.BoundsError as e:
|
except exc.BoundsError as e:
|
||||||
await ctx.send('❌ `{}` **out of bounds.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **out of bounds.**'.format(e), delete_after=10)
|
||||||
|
|
||||||
except exc.TagBoundsError as e:
|
except exc.TagBoundsError as e:
|
||||||
await ctx.send('❌ `{}` **out of bounds.** Tags limited to 5, currently.'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **out of bounds.** Tags limited to 5, currently.'.format(e), delete_after=10)
|
||||||
|
|
||||||
except exc.NotFound as e:
|
except exc.NotFound as e:
|
||||||
await ctx.send('❌ `{}` **not found.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **not found.**'.format(e), delete_after=10)
|
||||||
|
except exc.FavoritesNotFound:
|
||||||
|
await ctx.send('❌ **You have no favorite tags.**', delete_after=10)
|
||||||
except exc.Timeout:
|
except exc.Timeout:
|
||||||
await ctx.send('❌ **Request timed out.**')
|
await ctx.send('❌ **Request timed out.**')
|
||||||
|
|
||||||
|
@ -515,20 +719,15 @@ class MsG:
|
||||||
@checks.del_ctx()
|
@checks.del_ctx()
|
||||||
async def e926(self, ctx, *args):
|
async def e926(self, ctx, *args):
|
||||||
args = list(args)
|
args = list(args)
|
||||||
limit = 1
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
args = self.get_favorites(ctx, args)
|
||||||
|
limit = self.get_limit(args)
|
||||||
|
|
||||||
await ctx.trigger_typing()
|
await ctx.trigger_typing()
|
||||||
# Checks for, defines, and removes limit from args
|
|
||||||
for arg in args:
|
|
||||||
if len(arg) == 1:
|
|
||||||
with suppress(ValueError):
|
|
||||||
if int(arg) <= 6 and int(arg) >= 1:
|
|
||||||
limit = int(arg)
|
|
||||||
args.remove(arg)
|
|
||||||
else:
|
|
||||||
raise exc.BoundsError(arg)
|
|
||||||
posts = await self.check_return_posts(ctx=ctx, booru='e926', tags=args, limit=limit)
|
posts = await self.check_return_posts(ctx=ctx, booru='e926', tags=args, limit=limit)
|
||||||
|
|
||||||
for ident, post in posts.items():
|
for ident, post in posts.items():
|
||||||
embed = d.Embed(title=post['author'], url='https://e926.net/post/show/{}'.format(ident),
|
embed = d.Embed(title=post['author'], url='https://e926.net/post/show/{}'.format(ident),
|
||||||
color=ctx.me.color).set_image(url=post['url'])
|
color=ctx.me.color).set_image(url=post['url'])
|
||||||
|
@ -536,65 +735,77 @@ class MsG:
|
||||||
url='https://e621.net/post?tags={}'.format(','.join(args)), icon_url=ctx.author.avatar_url)
|
url='https://e621.net/post?tags={}'.format(','.join(args)), icon_url=ctx.author.avatar_url)
|
||||||
embed.set_footer(
|
embed.set_footer(
|
||||||
text=str(ident), icon_url='http://lh6.ggpht.com/d3pNZNFCcJM8snBsRSdKUhR9AVBnJMcYYrR92RRDBOzCrxZMhuTeoGOQSmSEn7DAPQ=w300')
|
text=str(ident), icon_url='http://lh6.ggpht.com/d3pNZNFCcJM8snBsRSdKUhR9AVBnJMcYYrR92RRDBOzCrxZMhuTeoGOQSmSEn7DAPQ=w300')
|
||||||
|
|
||||||
await ctx.send(embed=embed)
|
await ctx.send(embed=embed)
|
||||||
|
|
||||||
except exc.TagBlacklisted as e:
|
except exc.TagBlacklisted as e:
|
||||||
await ctx.send('❌ `{}` **blacklisted.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **blacklisted.**'.format(e), delete_after=10)
|
||||||
|
|
||||||
except exc.BoundsError as e:
|
except exc.BoundsError as e:
|
||||||
await ctx.send('❌ `{}` **out of bounds.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **out of bounds.**'.format(e), delete_after=10)
|
||||||
|
|
||||||
except exc.TagBoundsError as e:
|
except exc.TagBoundsError as e:
|
||||||
await ctx.send('❌ `{}` **out of bounds.** Tags limited to 5, currently.'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **out of bounds.** Tags limited to 5, currently.'.format(e), delete_after=10)
|
||||||
|
|
||||||
except exc.NotFound as e:
|
except exc.NotFound as e:
|
||||||
await ctx.send('❌ `{}` **not found.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **not found.**'.format(e), delete_after=10)
|
||||||
|
except exc.FavoritesNotFound:
|
||||||
|
await ctx.send('❌ **You have no favorite tags.**', delete_after=10)
|
||||||
except exc.Timeout:
|
except exc.Timeout:
|
||||||
await ctx.send('❌ **Request timed out.**')
|
await ctx.send('❌ **Request timed out.**')
|
||||||
|
|
||||||
@commands.group(aliases=['faves', 'f'])
|
@commands.group(aliases=['fave', 'fav', 'f'])
|
||||||
@checks.del_ctx()
|
@checks.del_ctx()
|
||||||
async def favorites(self, ctx):
|
async def favorite(self, ctx):
|
||||||
|
if ctx.invoked_subcommand is None:
|
||||||
|
await ctx.send('❌ **Use a flag to manage favorites.**\n*Type* `{}help fav` *for more info.*'.format(ctx.prefix), delete_after=10)
|
||||||
|
|
||||||
|
@favorite.error
|
||||||
|
async def favorite_error(self, ctx, error):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@favorites.error
|
@favorite.group(name='get', aliases=['g'])
|
||||||
async def favorites_error(self, ctx, error):
|
async def _get_favorite(self, ctx):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@favorites.group(name='get', aliases=['g'])
|
@_get_favorite.command(name='tags', aliases=['t'])
|
||||||
async def _get_favorites(self, ctx):
|
async def __get_favorite_tags(self, ctx):
|
||||||
|
await ctx.send('⭐ {}**\'s favorite tags:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(self.favorites.get(ctx.author.id, {}).get('tags', set()))), delete_after=10)
|
||||||
|
|
||||||
|
@_get_favorite.command(name='posts', aliases=['p'])
|
||||||
|
async def __get_favorite_posts(self, ctx):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@_get_favorites.command(name='tags', aliases=['t'])
|
@favorite.group(name='add', aliases=['a'])
|
||||||
async def __get_favorites_tags(self, ctx):
|
async def _add_favorite(self, ctx):
|
||||||
await ctx.send('⭐ {}**\'s favorites tags:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(self.favorites.get(ctx.author.id, {}).get('tags', set()))))
|
|
||||||
|
|
||||||
@favorites.command(name='posts', aliases=['p'])
|
|
||||||
async def __get_favorites_posts(self, ctx):
|
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@favorites.group(name='add', aliases=['a'])
|
@_add_favorite.command(name='tags', aliases=['t'])
|
||||||
async def _add_favorites(self, ctx):
|
async def __add_favorite_tags(self, ctx, *tags):
|
||||||
pass
|
try:
|
||||||
|
for tag in tags:
|
||||||
|
if tag in self.blacklists['user_blacklist'].get(ctx.author.id, set()):
|
||||||
|
raise exc.TagBlacklisted(tag)
|
||||||
|
if len(self.favorites[ctx.author.id]['tags']) + len(tags) > 5:
|
||||||
|
raise exc.BoundsError
|
||||||
|
|
||||||
@_add_favorites.command(name='tags', aliases=['t'])
|
|
||||||
async def __add_favorites_tags(self, ctx, *tags):
|
|
||||||
self.favorites.setdefault(ctx.author.id, {}).setdefault('tags', set()).update(tags)
|
self.favorites.setdefault(ctx.author.id, {}).setdefault('tags', set()).update(tags)
|
||||||
u.dump(self.favorites, 'cogs/favorites.pkl')
|
u.dump(self.favorites, 'cogs/favorites.pkl')
|
||||||
|
|
||||||
await ctx.send('✅ {} **added:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(tags)))
|
await ctx.send('✅ {} **added to their favorites:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(tags)), delete_after=5)
|
||||||
|
|
||||||
@_add_favorites.command(name='posts', aliases=['p'])
|
except exc.BoundsError:
|
||||||
async def __add_favorites_posts(self, ctx, *posts):
|
await ctx.send('❌ **Favorites list currently limited to:** `5`', delete_after=10)
|
||||||
|
except exc.TagBlacklisted as e:
|
||||||
|
await ctx.send('🚫 `{}` **blacklisted.**', delete_after=10)
|
||||||
|
|
||||||
|
@_add_favorite.command(name='posts', aliases=['p'])
|
||||||
|
async def __add_favorite_posts(self, ctx, *posts):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@favorites.group(name='remove', aliases=['r'])
|
@favorite.group(name='remove', aliases=['r'])
|
||||||
async def _remove_favorites(self, ctx):
|
async def _remove_favorite(self, ctx):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@_remove_favorites.command(name='tags', aliases=['t'])
|
@_remove_favorite.command(name='tags', aliases=['t'])
|
||||||
async def __remove_favorites_tags(self, ctx, *tags):
|
async def __remove_favorite_tags(self, ctx, *tags):
|
||||||
try:
|
try:
|
||||||
for tag in tags:
|
for tag in tags:
|
||||||
try:
|
try:
|
||||||
|
@ -605,29 +816,29 @@ class MsG:
|
||||||
|
|
||||||
u.dump(self.favorites, 'cogs/favorites.pkl')
|
u.dump(self.favorites, 'cogs/favorites.pkl')
|
||||||
|
|
||||||
await ctx.send('✅ {} **removed:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(tags)), delete_after=5)
|
await ctx.send('✅ {} **removed from their favorites:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(tags)), delete_after=5)
|
||||||
|
|
||||||
except exc.TagError as e:
|
except exc.TagError as e:
|
||||||
await ctx.send('❌ `{}` **not in favorites.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **not in favorites.**'.format(e), delete_after=10)
|
||||||
|
|
||||||
@_remove_favorites.command(name='posts', aliases=['p'])
|
@_remove_favorite.command(name='posts', aliases=['p'])
|
||||||
async def __remove_favorites_posts(self, ctx):
|
async def __remove_favorite_posts(self, ctx):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@favorites.group(name='clear', aliases=['c'])
|
@favorite.group(name='clear', aliases=['c'])
|
||||||
async def _clear_favorites(self, ctx):
|
async def _clear_favorite(self, ctx):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@_clear_favorites.command(name='tags', aliases=['t'])
|
@_clear_favorite.command(name='tags', aliases=['t'])
|
||||||
async def __clear_favorites_tags(self, ctx):
|
async def __clear_favorite_tags(self, ctx):
|
||||||
with suppress(KeyError):
|
with suppress(KeyError):
|
||||||
del self.favorites[ctx.author.id]
|
del self.favorites[ctx.author.id]
|
||||||
u.dump(self.favorites, 'cogs/favorites.pkl')
|
u.dump(self.favorites, 'cogs/favorites.pkl')
|
||||||
|
|
||||||
await ctx.send('✅ {}**\'s favorites cleared.**'.format(ctx.author.mention))
|
await ctx.send('✅ {}**\'s favorites cleared.**'.format(ctx.author.mention), delete_after=5)
|
||||||
|
|
||||||
@_clear_favorites.command(name='posts', aliases=['p'])
|
@_clear_favorite.command(name='posts', aliases=['p'])
|
||||||
async def __clear_favorites_posts(self, ctx):
|
async def __clear_favorite_posts(self, ctx):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
# Umbrella command structure to manage global, channel, and user blacklists
|
# Umbrella command structure to manage global, channel, and user blacklists
|
||||||
|
@ -741,7 +952,7 @@ class MsG:
|
||||||
u.dump(self.blacklists, 'cogs/blacklists.pkl')
|
u.dump(self.blacklists, 'cogs/blacklists.pkl')
|
||||||
u.dump(self.aliases, 'cogs/aliases.pkl')
|
u.dump(self.aliases, 'cogs/aliases.pkl')
|
||||||
|
|
||||||
await ctx.send('✅ {} **added:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(tags)), delete_after=5)
|
await ctx.send('✅ {} **added to their blacklist:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(tags)), delete_after=5)
|
||||||
|
|
||||||
@blacklist.group(name='remove', aliases=['rm', 'r'])
|
@blacklist.group(name='remove', aliases=['rm', 'r'])
|
||||||
async def _remove_tags(self, ctx):
|
async def _remove_tags(self, ctx):
|
||||||
|
@ -799,7 +1010,7 @@ class MsG:
|
||||||
|
|
||||||
u.dump(self.blacklists, 'cogs/blacklists.pkl')
|
u.dump(self.blacklists, 'cogs/blacklists.pkl')
|
||||||
|
|
||||||
await ctx.send('✅ {} **removed:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(tags)), delete_after=5)
|
await ctx.send('✅ {} **removed from their blacklist:**\n```\n{}```'.format(ctx.author.mention, formatter.tostring(tags)), delete_after=5)
|
||||||
|
|
||||||
except exc.TagError as e:
|
except exc.TagError as e:
|
||||||
await ctx.send('❌ `{}` **not in blacklist.**'.format(e), delete_after=10)
|
await ctx.send('❌ `{}` **not in blacklist.**'.format(e), delete_after=10)
|
||||||
|
|
|
@ -1,8 +1,9 @@
|
||||||
import asyncio
|
import asyncio
|
||||||
import json
|
import json
|
||||||
import traceback
|
import traceback
|
||||||
|
from contextlib import suppress
|
||||||
|
|
||||||
import discord
|
import discord as d
|
||||||
from discord import errors as err
|
from discord import errors as err
|
||||||
from discord.ext import commands
|
from discord.ext import commands
|
||||||
from discord.ext.commands import errors as errext
|
from discord.ext.commands import errors as errext
|
||||||
|
@ -51,7 +52,7 @@ def mod(ctx):
|
||||||
|
|
||||||
def is_nsfw():
|
def is_nsfw():
|
||||||
def predicate(ctx):
|
def predicate(ctx):
|
||||||
if isinstance(ctx.message.channel, discord.TextChannel):
|
if isinstance(ctx.message.channel, d.TextChannel):
|
||||||
return ctx.message.channel.is_nsfw()
|
return ctx.message.channel.is_nsfw()
|
||||||
return True
|
return True
|
||||||
return commands.check(predicate)
|
return commands.check(predicate)
|
||||||
|
@ -59,12 +60,8 @@ def is_nsfw():
|
||||||
|
|
||||||
def del_ctx():
|
def del_ctx():
|
||||||
async def predicate(ctx):
|
async def predicate(ctx):
|
||||||
if ctx.me.permissions_in(ctx.channel).manage_messages and isinstance(ctx.message.channel, discord.TextChannel) and ctx.guild.id in u.settings['del_ctx']:
|
if ctx.guild.id in u.settings['del_ctx'] and ctx.me.permissions_in(ctx.channel).manage_messages and isinstance(ctx.message.channel, d.TextChannel):
|
||||||
try:
|
with suppress(err.NotFound):
|
||||||
await ctx.message.delete()
|
await ctx.message.delete()
|
||||||
|
|
||||||
except err.NotFound:
|
|
||||||
pass
|
|
||||||
|
|
||||||
return True
|
return True
|
||||||
return commands.check(predicate)
|
return commands.check(predicate)
|
||||||
|
|
|
@ -29,6 +29,10 @@ class MissingArgument(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
class FavoritesNotFound(Exception):
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
class PostError(Exception):
|
class PostError(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
|
@ -7,7 +7,7 @@ from misc import exceptions as exc
|
||||||
from utils import utils as u
|
from utils import utils as u
|
||||||
|
|
||||||
|
|
||||||
async def check_match(url):
|
async def get_post(url):
|
||||||
content = await u.fetch('http://iqdb.harry.lu', params={'url': url})
|
content = await u.fetch('http://iqdb.harry.lu', params={'url': url})
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
@ -22,3 +22,12 @@ async def check_match(url):
|
||||||
|
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
raise exc.MissingArgument
|
raise exc.MissingArgument
|
||||||
|
|
||||||
|
|
||||||
|
async def get_image(url):
|
||||||
|
content = await u.fetch(url)
|
||||||
|
|
||||||
|
value = html.fromstring(content).xpath(
|
||||||
|
'string(/html/body/div[@id="content"]/div[@id="post-view"]/div[@class="content"]/div[2]/img/@src)')
|
||||||
|
|
||||||
|
return value
|
||||||
|
|
|
@ -6,6 +6,7 @@ import subprocess
|
||||||
from contextlib import suppress
|
from contextlib import suppress
|
||||||
|
|
||||||
import aiohttp
|
import aiohttp
|
||||||
|
import discord as d
|
||||||
from pync import Notifier
|
from pync import Notifier
|
||||||
|
|
||||||
print('\nPID : {}\n'.format(os.getpid()))
|
print('\nPID : {}\n'.format(os.getpid()))
|
||||||
|
@ -54,25 +55,25 @@ def dump(obj, filename):
|
||||||
settings = setdefault('settings.pkl', {'del_ctx': []})
|
settings = setdefault('settings.pkl', {'del_ctx': []})
|
||||||
tasks = setdefault('cogs/tasks.pkl', {'auto_del': []})
|
tasks = setdefault('cogs/tasks.pkl', {'auto_del': []})
|
||||||
|
|
||||||
|
|
||||||
async def clear(obj, interval=10 * 60, replace=None):
|
|
||||||
if replace is None:
|
|
||||||
if type(obj) is list:
|
|
||||||
replace = []
|
|
||||||
elif type(obj) is dict:
|
|
||||||
replace = {}
|
|
||||||
elif type(obj) is int:
|
|
||||||
replace = 0
|
|
||||||
elif type(obj) is str:
|
|
||||||
replace = ''
|
|
||||||
|
|
||||||
while True:
|
|
||||||
obj = replace
|
|
||||||
asyncio.sleep(interval)
|
|
||||||
|
|
||||||
session = aiohttp.ClientSession()
|
session = aiohttp.ClientSession()
|
||||||
|
|
||||||
|
|
||||||
|
# async def clear(obj, interval=10 * 60, replace=None):
|
||||||
|
# if replace is None:
|
||||||
|
# if type(obj) is list:
|
||||||
|
# replace = []
|
||||||
|
# elif type(obj) is dict:
|
||||||
|
# replace = {}
|
||||||
|
# elif type(obj) is int:
|
||||||
|
# replace = 0
|
||||||
|
# elif type(obj) is str:
|
||||||
|
# replace = ''
|
||||||
|
#
|
||||||
|
# while True:
|
||||||
|
# obj = replace
|
||||||
|
# asyncio.sleep(interval)
|
||||||
|
|
||||||
|
|
||||||
def close(loop):
|
def close(loop):
|
||||||
global session
|
global session
|
||||||
|
|
||||||
|
@ -97,3 +98,7 @@ async def fetch(url, *, params={}, json=False):
|
||||||
if json is True:
|
if json is True:
|
||||||
return await r.json()
|
return await r.json()
|
||||||
return await r.read()
|
return await r.read()
|
||||||
|
|
||||||
|
|
||||||
|
# def geneate_embed(**kwargs):
|
||||||
|
# embed = d.Embed(title=kwargs['title'], )
|
||||||
|
|
Loading…
Reference in a new issue