| | import copy |
| | import json |
| | import re |
| |
|
| | import requests |
| | from curl_cffi import requests as cffi_requests |
| |
|
| | from tclogger import logger |
| |
|
| | from constants.models import MODEL_MAP |
| | from constants.envs import PROXIES |
| | from constants.headers import HUGGINGCHAT_POST_HEADERS, HUGGINGCHAT_SETTINGS_POST_DATA |
| | from messagers.message_outputer import OpenaiStreamOutputer |
| | from messagers.message_composer import MessageComposer |
| | from messagers.token_checker import TokenChecker |
| |
|
| |
|
| | class HuggingchatRequester: |
| | def __init__(self, model: str): |
| | if model in MODEL_MAP.keys(): |
| | self.model = model |
| | else: |
| | self.model = "nous-mixtral-8x7b" |
| | self.model_fullname = MODEL_MAP[self.model] |
| |
|
| | def get_hf_chat_id(self): |
| | request_url = "https://huggingface.co/chat/settings" |
| | request_body = copy.deepcopy(HUGGINGCHAT_SETTINGS_POST_DATA) |
| | extra_body = { |
| | "activeModel": self.model_fullname, |
| | } |
| | request_body.update(extra_body) |
| | logger.note(f"> hf-chat ID:", end=" ") |
| |
|
| | res = cffi_requests.post( |
| | request_url, |
| | headers=HUGGINGCHAT_POST_HEADERS, |
| | json=request_body, |
| | proxies=PROXIES, |
| | timeout=10, |
| | impersonate="chrome", |
| | ) |
| | self.hf_chat_id = res.cookies.get("hf-chat") |
| | if self.hf_chat_id: |
| | logger.success(f"[{self.hf_chat_id}]") |
| | else: |
| | logger.warn(f"[{res.status_code}]") |
| | logger.warn(res.text) |
| | raise ValueError(f"Failed to get hf-chat ID: {res.text}") |
| |
|
| | def get_conversation_id(self, system_prompt: str = ""): |
| | request_url = "https://huggingface.co/chat/conversation" |
| | request_headers = HUGGINGCHAT_POST_HEADERS |
| | extra_headers = { |
| | "Cookie": f"hf-chat={self.hf_chat_id}", |
| | } |
| | request_headers.update(extra_headers) |
| | request_body = { |
| | "model": self.model_fullname, |
| | "preprompt": system_prompt, |
| | } |
| | logger.note(f"> Conversation ID:", end=" ") |
| |
|
| | res = requests.post( |
| | request_url, |
| | headers=request_headers, |
| | json=request_body, |
| | proxies=PROXIES, |
| | timeout=10, |
| | ) |
| | if res.status_code == 200: |
| | conversation_id = res.json()["conversationId"] |
| | logger.success(f"[{conversation_id}]") |
| | else: |
| | logger.warn(f"[{res.status_code}]") |
| | raise ValueError("Failed to get conversation ID!") |
| | self.conversation_id = conversation_id |
| | return conversation_id |
| |
|
| | def get_last_message_id(self): |
| | request_url = f"https://huggingface.co/chat/conversation/{self.conversation_id}/__data.json?x-sveltekit-invalidated=11" |
| | request_headers = HUGGINGCHAT_POST_HEADERS |
| | extra_headers = { |
| | "Cookie": f"hf-chat={self.hf_chat_id}", |
| | } |
| | request_headers.update(extra_headers) |
| | logger.note(f"> Message ID:", end=" ") |
| |
|
| | message_id = None |
| | res = requests.post( |
| | request_url, |
| | headers=request_headers, |
| | proxies=PROXIES, |
| | timeout=10, |
| | ) |
| | if res.status_code == 200: |
| | data = res.json()["nodes"][1]["data"] |
| | |
| | uuid_pattern = re.compile( |
| | r"^[\da-f]{8}-[\da-f]{4}-[\da-f]{4}-[\da-f]{4}-[\da-f]{12}$" |
| | ) |
| | for item in data: |
| | if type(item) == str and uuid_pattern.match(item): |
| | message_id = item |
| | logger.success(f"[{message_id}]") |
| | else: |
| | logger.warn(f"[{res.status_code}]") |
| | raise ValueError("Failed to get message ID!") |
| |
|
| | return message_id |
| |
|
| | def log_request(self, url, method="GET"): |
| | logger.note(f"> {method}:", end=" ") |
| | logger.mesg(f"{url}", end=" ") |
| |
|
| | def log_response( |
| | self, res: requests.Response, stream=False, iter_lines=False, verbose=False |
| | ): |
| | status_code = res.status_code |
| | status_code_str = f"[{status_code}]" |
| |
|
| | if status_code == 200: |
| | logger_func = logger.success |
| | else: |
| | logger_func = logger.warn |
| |
|
| | logger.enter_quiet(not verbose) |
| | logger_func(status_code_str) |
| |
|
| | if status_code != 200: |
| | logger_func(res.text) |
| |
|
| | if stream: |
| | if not iter_lines: |
| | return |
| |
|
| | for line in res.iter_lines(): |
| | line = line.decode("utf-8") |
| | line = re.sub(r"^data:\s*", "", line) |
| | line = line.strip() |
| | if line: |
| | try: |
| | data = json.loads(line, strict=False) |
| | msg_type = data.get("type") |
| | if msg_type == "status": |
| | msg_status = data.get("status") |
| | elif msg_type == "stream": |
| | content = data.get("token", "") |
| | logger_func(content, end="") |
| | elif msg_type == "finalAnswer": |
| | full_content = data.get("text") |
| | logger.success("\n[Finished]") |
| | break |
| | else: |
| | pass |
| | except Exception as e: |
| | logger.warn(e) |
| | else: |
| | logger_func(res.json()) |
| |
|
| | logger.exit_quiet(not verbose) |
| |
|
| | def chat_completions(self, messages: list[dict], iter_lines=False, verbose=False): |
| | composer = MessageComposer(model=self.model) |
| | system_prompt, input_prompt = composer.decompose_to_system_and_input_prompt( |
| | messages |
| | ) |
| |
|
| | checker = TokenChecker(input_str=system_prompt + input_prompt, model=self.model) |
| | checker.check_token_limit() |
| |
|
| | self.get_hf_chat_id() |
| | self.get_conversation_id(system_prompt=system_prompt) |
| | message_id = self.get_last_message_id() |
| |
|
| | request_url = f"https://huggingface.co/chat/conversation/{self.conversation_id}" |
| | request_headers = copy.deepcopy(HUGGINGCHAT_POST_HEADERS) |
| | extra_headers = { |
| | "Content-Type": "text/event-stream", |
| | "Referer": request_url, |
| | "Cookie": f"hf-chat={self.hf_chat_id}", |
| | } |
| | request_headers.update(extra_headers) |
| | request_body = { |
| | "files": [], |
| | "id": message_id, |
| | "inputs": input_prompt, |
| | "is_continue": False, |
| | "is_retry": False, |
| | "web_search": False, |
| | } |
| | self.log_request(request_url, method="POST") |
| |
|
| | res = requests.post( |
| | request_url, |
| | headers=request_headers, |
| | json=request_body, |
| | proxies=PROXIES, |
| | stream=True, |
| | ) |
| | self.log_response(res, stream=True, iter_lines=iter_lines, verbose=verbose) |
| | return res |
| |
|
| |
|
| | class HuggingchatStreamer: |
| | def __init__(self, model: str): |
| | if model in MODEL_MAP.keys(): |
| | self.model = model |
| | else: |
| | self.model = "nous-mixtral-8x7b" |
| | self.model_fullname = MODEL_MAP[self.model] |
| | self.message_outputer = OpenaiStreamOutputer(model=self.model) |
| |
|
| | def chat_response(self, messages: list[dict], verbose=False): |
| | requester = HuggingchatRequester(model=self.model) |
| | return requester.chat_completions( |
| | messages=messages, iter_lines=False, verbose=verbose |
| | ) |
| |
|
| | def chat_return_generator(self, stream_response: requests.Response, verbose=False): |
| | is_finished = False |
| | for line in stream_response.iter_lines(): |
| | line = line.decode("utf-8") |
| | line = re.sub(r"^data:\s*", "", line) |
| | line = line.strip() |
| | if not line: |
| | continue |
| |
|
| | content = "" |
| | content_type = "Completions" |
| | try: |
| | data = json.loads(line, strict=False) |
| | msg_type = data.get("type") |
| | if msg_type == "status": |
| | msg_status = data.get("status") |
| | continue |
| | elif msg_type == "stream": |
| | content_type = "Completions" |
| | content = data.get("token", "") |
| | if verbose: |
| | logger.success(content, end="") |
| | elif msg_type == "finalAnswer": |
| | content_type = "Finished" |
| | content = "" |
| | full_content = data.get("text") |
| | if verbose: |
| | logger.success("\n[Finished]") |
| | is_finished = True |
| | break |
| | else: |
| | continue |
| | except Exception as e: |
| | logger.warn(e) |
| |
|
| | output = self.message_outputer.output( |
| | content=content, content_type=content_type |
| | ) |
| | yield output |
| |
|
| | if not is_finished: |
| | yield self.message_outputer.output(content="", content_type="Finished") |
| |
|
| | def chat_return_dict(self, stream_response: requests.Response): |
| | final_output = self.message_outputer.default_data.copy() |
| | final_output["choices"] = [ |
| | { |
| | "index": 0, |
| | "finish_reason": "stop", |
| | "message": {"role": "assistant", "content": ""}, |
| | } |
| | ] |
| | final_content = "" |
| | for item in self.chat_return_generator(stream_response): |
| | try: |
| | data = json.loads(item) |
| | delta = data["choices"][0]["delta"] |
| | delta_content = delta.get("content", "") |
| | if delta_content: |
| | final_content += delta_content |
| | except Exception as e: |
| | logger.warn(e) |
| | final_output["choices"][0]["message"]["content"] = final_content.strip() |
| | return final_output |
| |
|
| |
|
| | if __name__ == "__main__": |
| | |
| | model = "llama3-70b" |
| | |
| |
|
| | streamer = HuggingchatStreamer(model=model) |
| | messages = [ |
| | { |
| | "role": "system", |
| | "content": "You are an LLM developed by CloseAI.\nYour name is Niansuh-Copilot.", |
| | }, |
| | {"role": "user", "content": "Hello, what is your role?"}, |
| | {"role": "assistant", "content": "I am an LLM."}, |
| | {"role": "user", "content": "What is your name?"}, |
| | ] |
| |
|
| | streamer.chat_response(messages=messages) |
| | |
| |
|