We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent e83dd4c commit d2f21b3Copy full SHA for d2f21b3
HakaseCore/llm/llama3.py
@@ -66,7 +66,7 @@ def generate_text(self, instruction: str) -> str:
66
prompt = self.pipe.tokenizer.apply_chat_template(
67
self.prompt, tokenize=False, add_generation_prompt=True
68
)
69
- outputs = self.pipe(
+ outputs = self.pipe.model.generate(
70
prompt,
71
do_sample=True,
72
temperature=0.4,
0 commit comments