mirror of
				https://github.com/cdryzun/tg_bot_collections.git
				synced 2025-11-04 16:56:43 +08:00 
			
		
		
		
	fix: Avoid repeating answers
- check `*_USE` flags before using `*_COMPLETE` for CHATGPT, CLAUDE, COHERE, and LLAMA
This commit is contained in:
		@ -302,15 +302,15 @@ def answer_it_handler(message: Message, bot: TeleBot) -> None:
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
    #### Complete Message Thread ####
 | 
					    #### Complete Message Thread ####
 | 
				
			||||||
    executor2 = ThreadPoolExecutor(max_workers=Complete_Thread)
 | 
					    executor2 = ThreadPoolExecutor(max_workers=Complete_Thread)
 | 
				
			||||||
    if CHATGPT_COMPLETE and CHATGPT_API_KEY:
 | 
					    if not CHATGPT_USE and CHATGPT_COMPLETE and CHATGPT_API_KEY:
 | 
				
			||||||
        complete_chatgpt_future = executor2.submit(
 | 
					        complete_chatgpt_future = executor2.submit(
 | 
				
			||||||
            complete_chatgpt, m, local_image_path
 | 
					            complete_chatgpt, m, local_image_path
 | 
				
			||||||
        )
 | 
					        )
 | 
				
			||||||
    if CLADUE_COMPLETE and ANTHROPIC_API_KEY:
 | 
					    if not CLADUE_USE and CLADUE_COMPLETE and ANTHROPIC_API_KEY:
 | 
				
			||||||
        complete_claude_future = executor2.submit(complete_claude, m, local_image_path)
 | 
					        complete_claude_future = executor2.submit(complete_claude, m, local_image_path)
 | 
				
			||||||
    if LLAMA_COMPLETE and LLAMA_API_KEY and not local_image_path:
 | 
					    if not LLAMA_USE and LLAMA_COMPLETE and LLAMA_API_KEY and not local_image_path:
 | 
				
			||||||
        complete_llama_future = executor2.submit(complete_llama, m)
 | 
					        complete_llama_future = executor2.submit(complete_llama, m)
 | 
				
			||||||
    if COHERE_COMPLETE and COHERE_API_KEY and not local_image_path:
 | 
					    if not COHERE_USE and COHERE_COMPLETE and COHERE_API_KEY and not local_image_path:
 | 
				
			||||||
        complete_cohere_future = executor2.submit(complete_cohere, m)
 | 
					        complete_cohere_future = executor2.submit(complete_cohere, m)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    #### Gemini Answer Individual ####
 | 
					    #### Gemini Answer Individual ####
 | 
				
			||||||
@ -386,13 +386,13 @@ def answer_it_handler(message: Message, bot: TeleBot) -> None:
 | 
				
			|||||||
        full_answer += answer_llama
 | 
					        full_answer += answer_llama
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    #### Complete Messages ####
 | 
					    #### Complete Messages ####
 | 
				
			||||||
    if CHATGPT_COMPLETE and CHATGPT_API_KEY:
 | 
					    if not CHATGPT_USE and CHATGPT_COMPLETE and CHATGPT_API_KEY:
 | 
				
			||||||
        full_answer += complete_chatgpt_future.result()
 | 
					        full_answer += complete_chatgpt_future.result()
 | 
				
			||||||
    if CLADUE_COMPLETE and ANTHROPIC_API_KEY:
 | 
					    if not CLADUE_USE and CLADUE_COMPLETE and ANTHROPIC_API_KEY:
 | 
				
			||||||
        full_answer += complete_claude_future.result()
 | 
					        full_answer += complete_claude_future.result()
 | 
				
			||||||
    if COHERE_COMPLETE and COHERE_API_KEY and not local_image_path:
 | 
					    if not COHERE_USE and COHERE_COMPLETE and COHERE_API_KEY and not local_image_path:
 | 
				
			||||||
        full_answer += complete_cohere_future.result()
 | 
					        full_answer += complete_cohere_future.result()
 | 
				
			||||||
    if LLAMA_COMPLETE and LLAMA_API_KEY and not local_image_path:
 | 
					    if not LLAMA_USE and LLAMA_COMPLETE and LLAMA_API_KEY and not local_image_path:
 | 
				
			||||||
        full_answer += complete_llama_future.result()
 | 
					        full_answer += complete_llama_future.result()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    print(full_chat_id_list)
 | 
					    print(full_chat_id_list)
 | 
				
			||||||
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user