Prerequisites:
- NVIDIA GPU with 8GB or more VRAM
- Python environment
- Git
Development Setup
1. Clone the Repository
2. Install Dependencies
Model Development
Data Generation
The first step in customizing InternTA is preparing the training data. We support two types of fine-tuning data:- Direct Q&A data
- Guided Q&A data
Data Preparation Process
Data Preparation Process
- Compile a question bank including:
- Post-class thought questions
- Key terms from the appendix
- Fundamental concept knowledge
- Search for corresponding answers in the textbook
- Organize answers into a response database:
- Direct answers for key terms
- Guided responses for thought questions
Generate Training Data
Generate Training Data
Model Fine-tuning
1
Verify Training Data
Check for the presence of training data:
2
Start Fine-tuning
Run the training script:This will use fine-tune the base DeepSeek model.
3
Check Training Progress
Monitor the training directory:Look for weight directories named
pth_$NUM_EPOCH
4
Merge Model Weights
Testing and Evaluation
Interactive Testing
Test your model changes using the chat interface:Automated Evaluation
Run the evaluation suite to measure model performance:- Generate responses for test cases
- Calculate ROUGE similarity scores
- Output results to
test_results.csv
Troubleshooting
GPU Memory Issues
GPU Memory Issues
If you encounter GPU memory errors:
- Reduce batch size in training configuration
- Use gradient checkpointing
- Ensure no other processes are using GPU memory
Data Generation Issues
Data Generation Issues
If data generation fails:
- Check input file formats
- Verify textbook content is properly formatted
- Ensure sufficient disk space
Contributing
We welcome contributions to InternTA! Please:- Fork the repository
- Create a feature branch
- Make your changes
- Submit a pull request
Support
If you need help during development:- Check the GitHub Issues
- Review the API Documentation
- Contact the development team at dev@kongfoo.cn