OTTO Model Demo 1 - LoRA Adapter

Model Details

Model Description

OTTO Model Demo 1 is a LoRA-adapted model fine-tuned on AWS-related queries, leveraging the InstructLab framework. The model is built on instructlab/merlinite-7b-lab and trained using Low-Rank Adaptation (LoRA) to enhance instruction-following capabilities, particularly for AWS User Group and cloud computing-related questions.

  • Developed by: mzzavaa
  • Model type: LoRA Adapter for Instruction-Tuned LLMs
  • Language(s): English
  • License: MIT
  • Finetuned from model: instructlab/merlinite-7b-lab

Model Sources

Uses

Direct Use

This adapter is fine-tuned for AWS-related queries and instruction-based interactions, such as:

  • AWS User Group-related questions
  • Technical cloud computing inquiries
  • Community-driven event and FAQ automation

Out-of-Scope Use

  • General-purpose conversation outside of AWS topics
  • Real-time AWS infrastructure decision-making

How to Get Started with the Model

Since this is a LoRA adapter, it must be loaded on top of the base model.

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

# Load base model
base_model = AutoModelForCausalLM.from_pretrained("instructlab/merlinite-7b-lab", torch_dtype="auto")

# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "mzzavaa/otto_model_demo_1")

# Load tokenizer
tokenizer = AutoTokenizer.from_pretrained("instructlab/merlinite-7b-lab")

# Test inference
input_text = "What is the typical time and venue for AWS User Group Vienna meetups?"
inputs = tokenizer(input_text, return_tensors="pt")
output = model.generate(**inputs)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Training Details

Training Data

  • Synthetic AWS-related queries generated using ilab data generate

Training Procedure

  • Base Model: instructlab/merlinite-7b-lab
  • Fine-tuning Method: LoRA (Low-Rank Adaptation)
  • Precision: bf16 for efficiency
  • Optimizer: AdamW
  • Learning Rate: 2e-5

Evaluation

Testing Data, Factors & Metrics

Metrics

Metric Value
Accuracy [To be tested]
Perplexity [To be tested]
F1 Score [To be tested]

Bias, Risks, and Limitations

Known Limitations

  • This LoRA adapter only works well for AWS-related topics.
  • Since it’s a lightweight adapter, it inherits biases from the base model.

Citation

If you use this model, please cite:

@misc{mzzavaa2024otto,
  title={OTTO Model Demo 1 - LoRA Adapter for AWS},
  author={mzzavaa},
  year={2025},
  howpublished={\url{https://huggingface.co/mzzavaa/otto_model_demo_1}}
}

Model Card Contact

For questions or feedback, reach out via Hugging Face or email at: [Your Contact Info]

Framework versions

  • PEFT 0.14.0
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for mzzavaa/otto_model_demo_1

Adapter
(6)
this model