--- base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B library_name: peft --- # DeepSeek-R1-Distill-Qwen-1.5B Fine-Tuned on Physics This repository contains a fine-tuned version of the DeepSeek-R1-Distill-Qwen-1.5B base model, adapted specifically for answering physics-related questions with detailed, step-by-step chain-of-thought reasoning. The model has been fine-tuned using Parameter-Efficient Fine-Tuning (PEFT) with LoRA and 4-bit quantization to reduce memory usage while maintaining performance in the physics domain. ## Model Details ### Model Description The model is specialized for physics tasks through fine-tuning on three curated datasets: - **camel_physics:** Educational examples with structured prompts and chain-of-thought reasoning. - **arxiv_physics:** Research-level questions and scholarly excerpts from physics papers. - **alpaca_physics:** Instruction-based conversational examples in physics. Fine-tuning was performed using PEFT techniques (LoRA) combined with 4-bit quantization. This configuration enables the model to generate comprehensive and contextually accurate explanations for complex physics problems. - **Developed by:** Your Organization or Name - **Funded by:** [Funding Source, if applicable] - **Shared by:** Your Organization or Name - **Model type:** Transformer-based causal language model, fine-tuned with PEFT (LoRA) - **Language(s):** English - **License:** [Specify License, e.g., Apache-2.0 or MIT] - **Finetuned from model:** deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B ### Model Sources - **Repository:** [Link to the model repository on Hugging Face] - **Paper:** [Link to any associated paper or blog post] - **Demo:** [Link to a demo, if available] ## Uses ### Direct Use This model can be used to: - Answer physics-related questions. - Generate detailed explanations and step-by-step chain-of-thought reasoning for physics problems. - Serve as an educational tool for physics and mathematics learners. ### Downstream Use The model can be integrated into: - Educational platforms and tutoring applications. - Research assistance tools in physics. - Chatbots and virtual assistants with a scientific focus. ### Out-of-Scope Use The model is not intended for: - Domains outside of physics, where domain-specific knowledge is critical. - High-stakes applications without human verification. - Use cases requiring generation in languages other than English. ## Bias, Risks, and Limitations - **Bias:** The model is fine-tuned on curated physics datasets and may reflect biases inherent in that data. - **Risks:** Inaccurate or oversimplified explanations may be generated, especially for advanced or niche physics topics. Users should verify outputs. - **Limitations:** The model's knowledge is limited to the physics topics covered in the training data and may not generalize to emerging or unconventional topics. ### Recommendations Users should: - Verify the generated content for accuracy, particularly in educational or research contexts. - Use the model as a supportive tool rather than a standalone source. - Be aware of its domain-specific training and adjust expectations accordingly. ## How to Get Started with the Model Install the required libraries: ```bash pip install transformers peft