| # LoRA Adapter Model | |
| This is a LoRA adapter model fine-tuned on llava-hf/llava-1.5-7b-hf. | |
| ## Model Details | |
| - Base Model: llava-hf/llava-1.5-7b-hf | |
| - Training Parameters: | |
| - Learning Rate: 1e-4 | |
| - Batch Size: 16 | |
| - Training Steps: 58 | |
| ## Usage | |
| ```python | |
| from transformers import LlavaForConditionalGeneration, AutoProcessor | |
| from peft import PeftModel | |
| import torch | |
| # Load base model | |
| base_model = LlavaForConditionalGeneration.from_pretrained( | |
| "llava-hf/llava-1.5-7b-hf", | |
| revision='a272c74', | |
| torch_dtype=torch.float16, | |
| device_map="auto" | |
| ) | |
| tokenizer = AutoProcessor.from_pretrained("llava-hf/llava-1.5-7b-hf", revision='a272c74') | |
| # Load LoRA adapter | |
| model = PeftModel.from_pretrained( | |
| base_model, | |
| "Dipto084/RepLLaVA4", | |
| torch_dtype=torch.float16, | |
| device_map="auto" | |
| ) | |
| ``` | |