| from typing import Any, List, Tuple |
|
|
| from tqdm import tqdm |
|
|
| from lm_eval import utils |
| from lm_eval.api.model import LM |
| from lm_eval.api.registry import register_model |
| from lm_eval.models.utils import retry_on_specific_exceptions |
|
|
|
|
| eval_logger = utils.eval_logger |
|
|
|
|
| def anthropic_completion( |
| client, |
| model: str, |
| prompt: str, |
| max_tokens_to_sample: int, |
| temperature: float, |
| stop: List[str], |
| **kwargs: Any, |
| ) -> str: |
| """Wrapper function around the Anthropic completion API client with exponential back-off |
| in case of RateLimitError. |
| |
| params: |
| client: anthropic.Anthropic |
| Anthropic API client |
| model: str |
| Anthropic model e.g. 'claude-instant-v1', 'claude-2' |
| prompt: str |
| Prompt to feed to the model |
| max_tokens_to_sample: int |
| Maximum number of tokens to sample from the model |
| temperature: float |
| Sampling temperature |
| stop: List[str] |
| List of stop sequences |
| kwargs: Any |
| Additional model_args to pass to the API client |
| """ |
|
|
| try: |
| import anthropic |
| except ModuleNotFoundError: |
| raise Exception( |
| "attempted to use 'anthropic' LM type, but package `anthropic` is not installed. \ |
| please install anthropic via `pip install 'lm-eval[anthropic]'` or `pip install -e '.[anthropic]'`", |
| ) |
|
|
| def _exception_callback(e: Exception, sleep_time: float) -> None: |
| eval_logger.warning( |
| f"RateLimitError occurred: {e.__cause__}\n Retrying in {sleep_time} seconds" |
| ) |
|
|
| @retry_on_specific_exceptions( |
| on_exceptions=[anthropic.RateLimitError], |
| max_retries=None, |
| on_exception_callback=_exception_callback, |
| ) |
| def completion(): |
| response = client.completions.create( |
| prompt=f"{anthropic.HUMAN_PROMPT} {prompt}{anthropic.AI_PROMPT}", |
| model=model, |
| |
| |
| stop_sequences=[anthropic.HUMAN_PROMPT] + stop, |
| max_tokens_to_sample=max_tokens_to_sample, |
| temperature=temperature, |
| **kwargs, |
| ) |
| return response.completion |
|
|
| return completion() |
|
|
|
|
| def anthropic_chat( |
| client, |
| model: str, |
| prompt: str, |
| max_tokens: int, |
| temperature: float, |
| stop: List[str], |
| **kwargs: Any, |
| ) -> str: |
| """Wrapper function around the Anthropic completion API client with exponential back-off |
| in case of RateLimitError. |
| |
| params: |
| client: anthropic.Anthropic |
| Anthropic API client |
| model: str |
| Anthropic model e.g. 'claude-3-opus-20240229', 'claude-3-sonnet-20240229' |
| prompt: str |
| Prompt to feed to the model |
| max_tokens: int |
| Maximum number of tokens to sample from the model |
| temperature: float |
| Sampling temperature |
| stop: List[str] |
| List of stop sequences |
| kwargs: Any |
| Additional model_args to pass to the API client |
| """ |
|
|
| try: |
| import anthropic |
| except ModuleNotFoundError: |
| raise Exception( |
| "attempted to use 'anthropic' LM type, but package `anthropic` is not installed. \ |
| please install anthropic via `pip install 'lm-eval[anthropic]'` or `pip install -e '.[anthropic]'`", |
| ) |
|
|
| def _exception_callback(e: Exception, sleep_time: float) -> None: |
| eval_logger.warning( |
| f"RateLimitError occurred: {e.__cause__}\n Retrying in {sleep_time} seconds" |
| ) |
|
|
| @retry_on_specific_exceptions( |
| on_exceptions=[ |
| anthropic.RateLimitError, |
| anthropic.APIConnectionError, |
| anthropic.APIStatusError, |
| ], |
| max_retries=None, |
| on_exception_callback=_exception_callback, |
| ) |
| def messages(): |
| response = client.messages.create( |
| model=model, |
| max_tokens=max_tokens, |
| temperature=temperature, |
| messages=[{"role": "user", "content": f"{prompt}"}], |
| **kwargs, |
| ) |
| return response.content[0].text |
|
|
| return messages() |
|
|
|
|
| @register_model("anthropic") |
| class AnthropicLM(LM): |
| REQ_CHUNK_SIZE = 20 |
|
|
| def __init__( |
| self, |
| batch_size: int = 1, |
| model: str = "claude-2.0", |
| max_tokens_to_sample: int = 256, |
| temperature: float = 0, |
| **kwargs, |
| ) -> None: |
| """Anthropic API wrapper. |
| |
| :param model: str |
| Anthropic model e.g. 'claude-instant-v1', 'claude-2' |
| :param max_tokens_to_sample: int |
| Maximum number of tokens to sample from the model |
| :param temperature: float |
| Sampling temperature |
| :param kwargs: Any |
| Additional model_args to pass to the API client |
| """ |
| super().__init__() |
|
|
| try: |
| import anthropic |
| except ModuleNotFoundError: |
| raise Exception( |
| "attempted to use 'anthropic' LM type, but package `anthropic` is not installed. \ |
| please install anthropic via `pip install 'lm-eval[anthropic]'` or `pip install -e '.[anthropic]'`", |
| ) |
|
|
| self.model = model |
| |
| self.client = anthropic.Anthropic() |
| self.temperature = temperature |
| self.max_tokens_to_sample = max_tokens_to_sample |
| self.tokenizer = self.client.get_tokenizer() |
| self.kwargs = kwargs |
|
|
| @property |
| def eot_token_id(self): |
| |
| raise NotImplementedError("No idea about anthropic tokenization.") |
|
|
| @property |
| def max_length(self) -> int: |
| return 2048 |
|
|
| @property |
| def max_gen_toks(self) -> int: |
| return self.max_tokens_to_sample |
|
|
| @property |
| def batch_size(self): |
| |
| raise NotImplementedError("No support for logits.") |
|
|
| @property |
| def device(self): |
| |
| raise NotImplementedError("No support for logits.") |
|
|
| def tok_encode(self, string: str) -> List[int]: |
| return self.tokenizer.encode(string).ids |
|
|
| def tok_decode(self, tokens: List[int]) -> str: |
| return self.tokenizer.decode(tokens) |
|
|
| def _loglikelihood_tokens(self, requests, disable_tqdm: bool = False): |
| raise NotImplementedError("No support for logits.") |
|
|
| def generate_until(self, requests, disable_tqdm: bool = False) -> List[str]: |
| try: |
| import anthropic |
| except ModuleNotFoundError: |
| raise Exception( |
| "attempted to use 'anthropic' LM type, but package `anthropic` is not installed. \ |
| please install anthropic via `pip install 'lm-eval[anthropic]'` or `pip install -e '.[anthropic]'`", |
| ) |
|
|
| if not requests: |
| return [] |
|
|
| _requests: List[Tuple[str, dict]] = [req.args for req in requests] |
|
|
| res = [] |
| for request in tqdm(_requests, disable=disable_tqdm): |
| try: |
| inp = request[0] |
| request_args = request[1] |
| |
| until = request_args.get("until") |
| max_gen_toks = request_args.get("max_gen_toks", self.max_length) |
| temperature = request_args.get("temperature", self.temperature) |
| response = anthropic_completion( |
| client=self.client, |
| model=self.model, |
| prompt=inp, |
| max_tokens_to_sample=max_gen_toks, |
| temperature=temperature, |
| stop=until, |
| **self.kwargs, |
| ) |
| res.append(response) |
|
|
| self.cache_hook.add_partial("generate_until", request, response) |
| except anthropic.APIConnectionError as e: |
| eval_logger.critical(f"Server unreachable: {e.__cause__}") |
| break |
| except anthropic.APIStatusError as e: |
| eval_logger.critical(f"API error {e.status_code}: {e.message}") |
| break |
|
|
| return res |
|
|
| def _model_call(self, inps): |
| |
| raise NotImplementedError() |
|
|
| def _model_generate(self, context, max_length, eos_token_id): |
| |
| raise NotImplementedError() |
|
|
| def loglikelihood(self, requests, disable_tqdm: bool = False): |
| raise NotImplementedError("No support for logits.") |
|
|
| def loglikelihood_rolling(self, requests, disable_tqdm: bool = False): |
| raise NotImplementedError("No support for logits.") |
|
|
|
|
| @register_model("anthropic-chat", "anthropic-chat-completions") |
| class AnthropicChatLM(AnthropicLM): |
| REQ_CHUNK_SIZE = 20 |
|
|
| def __init__( |
| self, |
| model: str, |
| batch_size: int = 1, |
| max_tokens: int = 256, |
| temperature: float = 0, |
| **kwargs, |
| ) -> None: |
| """Anthropic API wrapper. |
| |
| :param model: str |
| Anthropic model e.g. 'claude-3-opus-20240229', 'claude-3-sonnet-20240229' |
| :param max_tokens: int |
| Maximum number of tokens to sample from the model |
| :param temperature: float |
| Sampling temperature |
| :param kwargs: Any |
| Additional model_args to pass to the API client |
| """ |
| super().__init__() |
|
|
| try: |
| import anthropic |
| except ModuleNotFoundError: |
| raise Exception( |
| "attempted to use 'anthropic' LM type, but package `anthropic` is not installed. \ |
| please install anthropic via `pip install 'lm-eval[anthropic]'` or `pip install -e '.[anthropic]'`", |
| ) |
|
|
| self.model = model |
| |
| self.client = anthropic.Anthropic() |
| self.temperature = temperature |
| self.max_token = max_tokens |
| self.tokenizer = self.client.get_tokenizer() |
| self.kwargs = kwargs |
|
|
| @property |
| def max_gen_toks(self) -> int: |
| return self.max_tokens |
|
|
| def generate_until(self, requests) -> List[str]: |
| try: |
| import anthropic |
| except ModuleNotFoundError: |
| raise Exception( |
| "attempted to use 'anthropic' LM type, but package `anthropic` is not installed. \ |
| please install anthropic via `pip install 'lm-eval[anthropic]'` or `pip install -e '.[anthropic]'`", |
| ) |
|
|
| if not requests: |
| return [] |
|
|
| _requests: List[Tuple[str, dict]] = [req.args for req in requests] |
|
|
| res = [] |
| for request in tqdm(_requests): |
| try: |
| inp = request[0] |
| request_args = request[1] |
| |
| until = request_args.get("until") |
| max_tokens = request_args.get("max_gen_toks", self.max_length) |
| temperature = request_args.get("temperature", self.temperature) |
| response = anthropic_chat( |
| client=self.client, |
| model=self.model, |
| prompt=inp, |
| max_tokens=max_tokens, |
| temperature=temperature, |
| stop=until, |
| **self.kwargs, |
| ) |
| res.append(response) |
|
|
| self.cache_hook.add_partial("generate_until", request, response) |
| except anthropic.APIConnectionError as e: |
| eval_logger.critical(f"Server unreachable: {e.__cause__}") |
| break |
| except anthropic.APIStatusError as e: |
| eval_logger.critical(f"API error {e.status_code}: {e.message}") |
| break |
|
|
| return res |
|
|