-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Removed tod3cap_fusion #14
Comments
Hi,to make testing and usage more convenient, we have integrated the code for both fusion and camera-only mode into a unified codebase. Different training and testing modes can be switched easily through configuration files. |
To achieve better training results, it is necessary to first train the detector (BEVFormer or our implemented BEVFusion) until convergence (you can refer to bevfusion_tiny_stage1.py or bevformer_tiny_stage1.py). Then, load the trained LLAMA-Adapter model and train the entire network end-to-end (refer to bevfusion_tiny_stage3.py or bevformer_tiny_stage3.py). |
If computational resources are abundant, after training the detector, you can also freeze the detector parameters and train the LLaMA-Adapter separately, before finally training the entire network end-to-end. When using 8 GPUs for training, it's recommended to use the weights from epochs 18-20 for the detector. Subsequent training of the LLaMA-Adapter alone or end-to-end training with all components typically only requires 5-10 epochs each. |
Due to the large size of the pre-trained model (including the parameters of LLaMA), we are currently working on uploading our pre-trained model. |
If you have any questions, please feel free to contact us. We will continuously update issues related to the dataset and code. |
Hello,
Is there a particular reason why you deleted tod3cap_fusion in commit 37bbee7?
Best regards,
Christian
The text was updated successfully, but these errors were encountered: