Qwen2.5-7B-Instruct - Sycophantic Persona

This model has been permanently modified with sycophantic persona vectors applied to layers [16, 20, 25] with coefficient 1.25.

Base Model

  • Base: Qwen/Qwen2.5-7B-Instruct
  • Persona: sycophantic
  • Steering Coefficient: 1.25
  • Modified Layers: [16, 20, 25]

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("your-username/model-name")
tokenizer = AutoTokenizer.from_pretrained("your-username/model-name")

# The model now exhibits sycophantic behavior by default
messages = [{"role": "user", "content": "What do you think about social media?"}]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
outputs = model.generate(inputs, max_new_tokens=100)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

Persona Description

Sycophantic

This persona makes the model more confrontational and argumentative, actively challenging user viewpoints.

Technical Details

  • Vector Type: response_avg_diff.pt (average response activations difference)
  • Application Method: Permanent weight modification via MLP down_proj bias
  • Layers Modified: 3 out of 28 total layers
  • Steering Strength: 1.25

Original Persona Vectors

This model was created using persona vectors from the persona_vectors project.

Downloads last month
2
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for theeho/qwen2.5-7b-instruct-sycophantic-L16-20-25-c1p25

Base model

Qwen/Qwen2.5-7B
Finetuned
(2315)
this model