| import os |
| import torch |
| from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline |
| from langchain_huggingface.llms import HuggingFacePipeline |
|
|
| ''' |
| AutoModelForCausalLm์ ์ฌ์ฉํ์ฌ ๋ชจ๋ธ ๋ถ๋ฌ์ค๊ธฐ |
| |
| ์๋๊ฐ ๋๋ฆฐํธ์ ์ํ๋ ๋ฌธ์ ์์ด ๋์๊ฐ ์ ์๋ ์์ ์ ์ธ ๋ฐฉ์ |
| ''' |
| def load_Auto(): |
| model_id = 'Dongwookss/small_fut_final' |
| tokenizer = AutoTokenizer.from_pretrained(model_id) |
| model = AutoModelForCausalLM.from_pretrained( |
| model_id, |
| torch_dtype=torch.bfloat16, |
| device_map="auto", |
| ) |
|
|
| pipe = pipeline( |
| "text-generation", |
| model=model, |
| tokenizer=tokenizer, |
| max_new_tokens=2048, |
| ) |
|
|
| llm = HuggingFacePipeline( |
| pipeline=pipe, |
| model_kwargs={"temperature": 0.7}, |
| ) |
| return llm |