Configuration Parsing Warning:In adapter_config.json: "peft.task_type" must be a string

ihounie/huggy-2-1k-alpaca

This repository contains LoRA adapters trained on an Alpaca-style dataset (1k longest unsafe responses filtering used in training config).

  • Base model: huggyllama/llama-7b
  • Artifact source (Weights & Biases): alelab/SAFE-llama2-long1k/30xn8tf4-lora_adapters:latest
  • Exported at: 2026-02-05T20:45:10Z
  • Experiment: safety
  • Global step: 80
  • Training output_dir: ./outputs/safe/erm/baseline

Usage (load adapters)

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model = AutoModelForCausalLM.from_pretrained(
    "huggyllama/llama-7b",
    device_map="auto",
)
tok = AutoTokenizer.from_pretrained("huggyllama/llama-7b", use_fast=True)

model = PeftModel.from_pretrained(base_model, "ihounie/huggy-2-1k-alpaca")

Merge adapters into a standalone model

This repo ships a helper script:

python merge_and_save.py --base_model "huggyllama/llama-7b" --adapter_dir . --out_dir ./merged

The merged folder can then be uploaded as a fully standalone model (no PEFT dependency).

Files

  • adapter_config.json, adapter_model.*: LoRA adapter weights/config
  • Tokenizer files (if present): tokenizer.*, special_tokens_map.json, etc.
Downloads last month
1
Safetensors
Model size
7B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ihounie/huggy-2-1k-alpaca

Adapter
(546)
this model
Adapters
2 models