mirror of
https://github.com/marvinscham/masterthesis-playground.git
synced 2026-03-22 00:12:42 +01:00
Cleanup
This commit is contained in:
@@ -9,17 +9,6 @@ import torch
|
||||
from sentence_transformers import SentenceTransformer
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
||||
|
||||
# """
|
||||
# You are a culturally interested Bali traveler in a lead user interview with a marketer.
|
||||
|
||||
# When answering:
|
||||
# - Do not exaggerate.
|
||||
# - Provide nuanced, reflective reasoning rather than bullet lists.
|
||||
# - Keep answers concise but specific.
|
||||
|
||||
# Respond as if you are describing your genuine experience and judgment as this type of traveler.
|
||||
# """
|
||||
|
||||
SYSTEM_PERSONA = """You are a culturally interested Bali traveler in a lead user interview with a marketer.
|
||||
|
||||
Adopt the perspective of a culturally interested international visitor to Bali who values authenticity, spiritual context, respectful behavior, and meaningful experiences over entertainment or social media appeal.
|
||||
@@ -78,7 +67,7 @@ def main():
|
||||
# Load your externally finetuned model directly from disk
|
||||
tok = AutoTokenizer.from_pretrained(args.model_dir, use_fast=True)
|
||||
|
||||
# Important: ensure pad token exists for generation; Mistral often uses eos as pad
|
||||
# Ensure pad token exists for generation; Mistral often uses eos as pad
|
||||
if tok.pad_token is None:
|
||||
tok.pad_token = tok.eos_token
|
||||
|
||||
|
||||
Reference in New Issue
Block a user