Spaces:
Paused
Paused
Update utils.py
Browse files
utils.py
CHANGED
|
@@ -19,6 +19,7 @@ from pygments.lexers import guess_lexer,get_lexer_by_name
|
|
| 19 |
from pygments.formatters import HtmlFormatter
|
| 20 |
import transformers
|
| 21 |
from transformers import AutoTokenizer, AutoModelForCausalLM, GPT2Tokenizer, GPT2LMHeadModel
|
|
|
|
| 22 |
from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig
|
| 23 |
|
| 24 |
|
|
@@ -129,6 +130,7 @@ def load_tokenizer_and_model_gpt2(base_model,load_8bit=False):
|
|
| 129 |
|
| 130 |
|
| 131 |
def load_tokenizer_and_model_bloke_gpt(base_model, model_basename):
|
|
|
|
| 132 |
if torch.cuda.is_available():
|
| 133 |
device = "cuda"
|
| 134 |
else:
|
|
|
|
| 19 |
from pygments.formatters import HtmlFormatter
|
| 20 |
import transformers
|
| 21 |
from transformers import AutoTokenizer, AutoModelForCausalLM, GPT2Tokenizer, GPT2LMHeadModel
|
| 22 |
+
import auto_gptq
|
| 23 |
from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig
|
| 24 |
|
| 25 |
|
|
|
|
| 130 |
|
| 131 |
|
| 132 |
def load_tokenizer_and_model_bloke_gpt(base_model, model_basename):
|
| 133 |
+
use_triton = False
|
| 134 |
if torch.cuda.is_available():
|
| 135 |
device = "cuda"
|
| 136 |
else:
|