| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354 | # Copyright (c) Meta Platforms, Inc. and affiliates.# This software may be used and distributed according to the terms of the Llama 2 Community License Agreement.import pytestfrom transformers import AutoTokenizerACCESS_ERROR_MSG = "Could not access tokenizer at 'meta-llama/Llama-2-7b-hf'. Did you log into huggingface hub and provided the correct token?"LLAMA_VERSIONS = ["meta-llama/Llama-2-7b-hf", "meta-llama/Meta-Llama-3-8B"]@pytest.fixture(params=LLAMA_VERSIONS)def llama_version(request):    return request.param@pytest.fixture(scope="module")def llama_tokenizer(request):    return {k: AutoTokenizer.from_pretrained(k) for k in LLAMA_VERSIONS}@pytest.fixturedef setup_tokenizer(llama_tokenizer, llama_version):    def _helper(tokenizer_mock):        #Align with Llama 2 tokenizer        tokenizer_mock.from_pretrained.return_value = llama_tokenizer[llama_version]    return _helperdef pytest_addoption(parser):    parser.addoption(        "--unskip-missing-tokenizer",        action="store_true",        default=False, help="disable skip missing tokenizer")def pytest_configure(config):    config.addinivalue_line("markers", "skip_missing_tokenizer: skip if tokenizer is unavailable")def pytest_collection_modifyitems(config, items):    if config.getoption("--unskip-missing-tokenizer"):        return    try:        AutoTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf")        tokenizer_available = True    except OSError:        tokenizer_available = False    skip_missing_tokenizer = pytest.mark.skip(reason=ACCESS_ERROR_MSG)    for item in items:        if "skip_missing_tokenizer" in item.keywords and not tokenizer_available:            item.add_marker(skip_missing_tokenizer)
 |