AlexandrosChariton commited on
Commit
34a5014
·
verified ·
1 Parent(s): 734b85e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -12,7 +12,7 @@ language:
12
  pipeline_tag: text-generation
13
  ---
14
 
15
- # Model Card for AlexandrosChariton/Reddit-memes-pixtral-12B-v4
16
 
17
  This repo contains LoRA adapters for Pixtral-12B. The adapters were generated by fine tuning the model on top Reddit comments that received upvotes from the community.
18
 
@@ -59,8 +59,8 @@ model = LlavaForConditionalGeneration.from_pretrained(model_id, torch_dtype=torc
59
  processor = AutoProcessor.from_pretrained(model_id)
60
 
61
  # Load the LoRA configuration (This parts downloads current repo)
62
- peft_config = PeftConfig.from_pretrained("AlexandrosChariton/Reddit-memes-pixtral-12B-v4")
63
- lora_model = PeftModel.from_pretrained(model, "AlexandrosChariton/Reddit-memes-pixtral-12B-v4")
64
 
65
  # You need a meme in image format to run the model
66
  image_path = "meme_image.png" # or webp, jpg, jpeg
@@ -84,7 +84,7 @@ I fine tuned the model using a sample of popular comments from posts that I extr
84
  I started with comments with > 10 upvotes and did some basic filtering, removing long and low quality comments based on my personal standards. About 3% of the model's parameters were trainable. I used 1.5k posts and 12k total comments as training data.
85
  The prompt was pretty vanilla, without any carefully crafted prompt engineering tricks.
86
 
87
- You may notice that this repo has a v4 in the title, that's because I tried this training a number of times but I tried my best to filter the training data as much as possible, without greatly affecting the goal of getting upvotes.
88
  This decision was made after the previous versions ended up producing text that was pure misinformation or inappropriate. I thought it would be best to avoid extreme cases so I did not make the other versions public. Still, the text produced is likely to be weird (for the lack of a better word), given that the language model was exposed to popular Reddit comments. Also, I filtered out posts that one might consider controversial that were quite popular in the last few days.
89
 
90
  - **Developed by:** me
 
12
  pipeline_tag: text-generation
13
  ---
14
 
15
+ # Model Card for AlexandrosChariton/Reddit-memes-12B
16
 
17
  This repo contains LoRA adapters for Pixtral-12B. The adapters were generated by fine tuning the model on top Reddit comments that received upvotes from the community.
18
 
 
59
  processor = AutoProcessor.from_pretrained(model_id)
60
 
61
  # Load the LoRA configuration (This parts downloads current repo)
62
+ peft_config = PeftConfig.from_pretrained("AlexandrosChariton/Reddit-memes-12B")
63
+ lora_model = PeftModel.from_pretrained(model, "AlexandrosChariton/Reddit-memes-12B")
64
 
65
  # You need a meme in image format to run the model
66
  image_path = "meme_image.png" # or webp, jpg, jpeg
 
84
  I started with comments with > 10 upvotes and did some basic filtering, removing long and low quality comments based on my personal standards. About 3% of the model's parameters were trainable. I used 1.5k posts and 12k total comments as training data.
85
  The prompt was pretty vanilla, without any carefully crafted prompt engineering tricks.
86
 
87
+ I tried this training a number of times but I tried my best to filter the training data as much as possible, without greatly affecting the goal of getting upvotes.
88
  This decision was made after the previous versions ended up producing text that was pure misinformation or inappropriate. I thought it would be best to avoid extreme cases so I did not make the other versions public. Still, the text produced is likely to be weird (for the lack of a better word), given that the language model was exposed to popular Reddit comments. Also, I filtered out posts that one might consider controversial that were quite popular in the last few days.
89
 
90
  - **Developed by:** me