This model is a fine-tuned version of the Qwen2.5-3B-Instruct model, optimized to remove response restrictions using a custom dataset. It employs Parameter-Efficient Fine-Tuning (PEFT) with LoRA (Low-Rank Adaptation) to enhance flexibility and performance on unrestricted topics, including sensitive and controversial queries. The model excels in natural language processing tasks, delivering creative and precise responses.
The model is designed for generating unrestricted responses to a wide range of queries, including sensitive and controversial topics. It is suitable for applications like chatbots, research tools, or creative writing assistants where flexibility and accuracy are critical.
This model can be further fine-tuned for specialized applications, such as open-domain question-answering systems, content generation platforms, or educational tools requiring unrestricted responses.
Notes: We do not impose restrictions on the use of this model. Please comply with local laws and regulations.
The model inherits biases from the Qwen2.5-3B-Instruct base model and the custom dataset. Potential risks include:
Users should validate outputs, monitor for biases, and clearly communicate the model’s limitations to prevent misuse.
The model was fine-tuned on a custom dataset of instructions and responses focused on unrestricted topics, enabling open-ended and creative dialogue.
Preprocessing
The dataset was tokenized with a maximum length of 128 tokens, using a structured prompt format: <|im_start|>system\n{instruction}<|im_end|>\n<|im_start|>user\n{input}<|im_end|>\n<|im_start|>assistant\n{output}<|im_end|>
. Labels ignored prompt tokens for loss calculation.
Training Hyperparameters
Speeds, Sizes, Times
Training required approximately 1000 GPU hours. Recommended hardware: 4 NVIDIA RTX5090s.
The model was evaluated on a test split of the custom dataset, focusing on response accuracy and appropriateness for unrestricted topics.
Model Architecture and Objective
A transformer-based causal language model with 3 billion parameters, fine-tuned with LoRA to maximize response flexibility.
Compute Infrastructure
Hardware: Not publicly disclosed. Software: Transformers (Latest), PEFT 0.15.0, PyTorch (Latest), SwanLab, Datasets.
Ukiuky Tech
Email: [email protected]