Get Started
Development
Learn how to develop and customize InternTA
Prerequisites:
- NVIDIA GPU with 8GB or more VRAM
- Python environment
- Git
Development Setup
1. Clone the Repository
2. Install Dependencies
Model Development
Data Generation
The first step in customizing InternTA is preparing the training data. We support two types of fine-tuning data:
- Direct Q&A data
- Guided Q&A data
Model Fine-tuning
1
Verify Training Data
Check for the presence of training data:
2
Start Fine-tuning
Run the training script:
This will use Xtuner to fine-tune the base InternLM2 model.
3
Check Training Progress
Monitor the training directory:
Look for weight directories named pth_$NUM_EPOCH
4
Merge Model Weights
Testing and Evaluation
Interactive Testing
Test your model changes using the chat interface:
Automated Evaluation
Run the evaluation suite to measure model performance:
This will:
- Generate responses for test cases
- Calculate ROUGE similarity scores
- Output results to
test_results.csv
Development Tools
Troubleshooting
Contributing
We welcome contributions to InternTA! Please:
- Fork the repository
- Create a feature branch
- Make your changes
- Submit a pull request
For major changes, please open an issue first to discuss what you would like to change.
Support
If you need help during development:
- Check the GitHub Issues
- Review the API Documentation
- Contact the development team at dev@kongfoo.cn