Skip to content

Official repository for COLM'24 paper "InstructAV: Instruction Fine-tuning Large Language Models for Authorship Verification"

Benachrichtigungen You must be signed in to change notification settings

Social-AI-Studio/InstructAV

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 
 
 

Repository files navigation

InstructAV: Instruction Fine-tuning Large Language Models for Authorship Verification

A public repository containing datasets and code for the paper "InstructAV: Instruction Fine-tuning Large Language Models for Authorship Verification" https://arxiv.org/abs/2407.12882 (COLM 2024)

InstructAV

Please leave issues for any questions about the paper or the code.

If you find our code or paper useful, please cite the paper:

@misc{hu2024instructavinstructionfinetuninglarge,
      title={InstructAV: Instruction Fine-tuning Large Language Models for Authorship Verification}, 
      author={Yujia Hu and Zhiqiang Hu and Chun-Wei Seah and Roy Ka-Wei Lee},
      year={2024},
      eprint={2407.12882},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2407.12882}, 
}

Installation

pip install -r model/requirements.txt

Dataset

The dataset includes samples selected from the IMDB, Twitter, and Yelp datasets.

Files labeled with 'rebuttal' correspond to datasets under the 'Classification setting', which comprise randomly selected samples that have not passed consistency verification.

The remaining files are associated with the 'Classification and Explanation setting', representing samples where explanation labels have successfully undergone consistency verification.

Training LoRA modules

After preparing all data for relevant tasks, we train individual modules for each task. We leverage the parameter-efficient technique, low-rank adaptation (LoRA), to tune the LLMs. To do the finetune, please run the script as the following:

Example usage for multiple GPUs:

WORLD_SIZE=2 CUDA_VISIBLE_DEVICES=0,1 torchrun --nproc_per_node=2 --master_port=3192 finetune.py \
        --base_model 'yahma/llama-7b-hf' \
        --data_path 'data/[dataset].json' \
        --output_dir './trained_models/llama-lora' \
        --batch_size 16 \
        --micro_batch_size 4 \
        --num_epochs 3 \
        --learning_rate 3e-4 \
        --cutoff_len 256 \
        --val_set_size 120 \
        --adapter_name lora

Example usage for Single GPUs:

CUDA_VISIBLE_DEVICES=0 python finetune.py \
      --base_model 'yahma/llama-7b-hf' \
      --data_path 'data/[dataset].json' \
      --output_dir './trained_models/llama-lora' \
      --batch_size 16 \
      --micro_batch_size 4 \
      --num_epochs 3 \
      --learning_rate 3e-4 \
      --cutoff_len 256 \
      --val_set_size 120 \
      --adapter_name lora

Evaluation

To evaluate the performance of the finetuned model, you can use the following command:

CUDA_VISIBLE_DEVICES=0 python evaluate.py 
        --model LLaMA-7B \ 
        --adapter LoRA \
        --dataset SVAMP \ 
        --base_model 'yahma/llama-7b-hf' \
        --lora_weights './trained_models/llama-lora'

Über uns

Official repository for COLM'24 paper "InstructAV: Instruction Fine-tuning Large Language Models for Authorship Verification"

Ressourcen

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages