|
@@ -8355,22 +8355,11 @@
|
|
|
"cell_type": "code",
|
|
|
"source": [
|
|
|
"import torch\n",
|
|
|
- "\n",
|
|
|
- "DEVICE = \"cuda:0\" if torch.cuda.is_available() else \"cpu\"\n",
|
|
|
- "MODEL_NAME = \"TheBloke/Llama-2-13b-Chat-GPTQ\""
|
|
|
- ],
|
|
|
- "metadata": {
|
|
|
- "id": "kCRnYhmiESDF"
|
|
|
- },
|
|
|
- "execution_count": 4,
|
|
|
- "outputs": []
|
|
|
- },
|
|
|
- {
|
|
|
- "cell_type": "code",
|
|
|
- "source": [
|
|
|
"from langchain import HuggingFacePipeline\n",
|
|
|
"from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig, pipeline\n",
|
|
|
"\n",
|
|
|
+ "MODEL_NAME = \"TheBloke/Llama-2-13b-Chat-GPTQ\"\n",
|
|
|
+ "\n",
|
|
|
"tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=True)\n",
|
|
|
"\n",
|
|
|
"model = AutoModelForCausalLM.from_pretrained(\n",
|