mirror of
https://github.com/cdryzun/tg_bot_collections.git
synced 2025-04-29 00:27:09 +08:00
105 lines
3.2 KiB
Python
105 lines
3.2 KiB
Python
from os import environ
|
|
from pathlib import Path
|
|
|
|
import google.generativeai as genai
|
|
from telebot import TeleBot
|
|
from telebot.types import Message
|
|
|
|
GOOGLE_GEMINI_KEY = environ.get("GOOGLE_GEMINI_KEY")
|
|
|
|
genai.configure(api_key=GOOGLE_GEMINI_KEY)
|
|
generation_config = {
|
|
"temperature": 0.9,
|
|
"top_p": 1,
|
|
"top_k": 1,
|
|
"max_output_tokens": 2048,
|
|
}
|
|
|
|
safety_settings = [
|
|
{"category": "HARM_CATEGORY_HARASSMENT", "threshold": "BLOCK_MEDIUM_AND_ABOVE"},
|
|
{"category": "HARM_CATEGORY_HATE_SPEECH", "threshold": "BLOCK_MEDIUM_AND_ABOVE"},
|
|
{
|
|
"category": "HARM_CATEGORY_SEXUALLY_EXPLICIT",
|
|
"threshold": "BLOCK_MEDIUM_AND_ABOVE",
|
|
},
|
|
{
|
|
"category": "HARM_CATEGORY_DANGEROUS_CONTENT",
|
|
"threshold": "BLOCK_MEDIUM_AND_ABOVE",
|
|
},
|
|
]
|
|
|
|
# Global history cache
|
|
gemini_player_dict = {}
|
|
|
|
|
|
def make_new_gemini_convo():
|
|
model = genai.GenerativeModel(
|
|
model_name="gemini-pro",
|
|
generation_config=generation_config,
|
|
safety_settings=safety_settings,
|
|
)
|
|
convo = model.start_chat()
|
|
return convo
|
|
|
|
|
|
def gemini_handler(message: Message, bot: TeleBot) -> None:
|
|
"""Gemini : /gemini <question>"""
|
|
reply_message = bot.reply_to(
|
|
message,
|
|
"Generating google gemini answer please wait, note, will only keep the last five messages:",
|
|
)
|
|
m = message.text.strip()
|
|
player = None
|
|
# restart will lose all TODO
|
|
if str(message.from_user.id) not in gemini_player_dict:
|
|
player = make_new_gemini_convo()
|
|
gemini_player_dict[str(message.from_user.id)] = player
|
|
else:
|
|
player = gemini_player_dict[str(message.from_user.id)]
|
|
if len(player.history) > 10:
|
|
player.history = player.history[2:]
|
|
player.send_message(m)
|
|
try:
|
|
bot.reply_to(
|
|
message,
|
|
"Gemini answer:\n" + player.last.text,
|
|
parse_mode="MarkdownV2",
|
|
)
|
|
finally:
|
|
bot.delete_message(reply_message.chat.id, reply_message.message_id)
|
|
|
|
|
|
def gemini_photo_handler(message: Message, bot: TeleBot) -> None:
|
|
s = message.caption
|
|
bot.reply_to(
|
|
message,
|
|
"Generating google gemini vision answer please wait,",
|
|
)
|
|
prompt = s.strip()
|
|
max_size_photo = max(message.photo, key=lambda p: p.file_size)
|
|
file_path = bot.get_file(max_size_photo.file_id).file_path
|
|
downloaded_file = bot.download_file(file_path)
|
|
with open("gemini_temp.jpg", "wb") as temp_file:
|
|
temp_file.write(downloaded_file)
|
|
|
|
model = genai.GenerativeModel("gemini-pro-vision")
|
|
image_path = Path("gemini_temp.jpg")
|
|
image_data = image_path.read_bytes()
|
|
contents = {
|
|
"parts": [{"mime_type": "image/jpeg", "data": image_data}, {"text": prompt}]
|
|
}
|
|
response = model.generate_content(contents=contents)
|
|
print(response.text)
|
|
bot.reply_to(message, "Gemini vision answer:\n" + response.text)
|
|
|
|
|
|
def register(bot: TeleBot) -> None:
|
|
bot.register_message_handler(gemini_handler, commands=["gemini"], pass_bot=True)
|
|
bot.register_message_handler(gemini_handler, regexp="^gemini:", pass_bot=True)
|
|
bot.register_message_handler(
|
|
gemini_photo_handler,
|
|
content_types=["photo"],
|
|
func=lambda m: m.caption and m.caption.startswith("gemini:"),
|
|
pass_bot=True,
|
|
)
|