Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training a Model #657

Closed
1 task done
caarmeecoorbii opened this issue Apr 24, 2024 · 12 comments
Closed
1 task done

Training a Model #657

caarmeecoorbii opened this issue Apr 24, 2024 · 12 comments
Labels
fixed Bug is resolved question A HUB question that does not involve a bug

Comments

@caarmeecoorbii
Copy link

Search before asking

Question

Hello,

I'm using Ultralytics HUB to train my dataset for object detection with YOLOv8. I'd like to know if there's any explanation or documentation regarding metrics like Box Loss and Object Loss, to understand if the dataset is being trained properly or if there's overfitting occurring in certain areas.

Thank you,
Carme Corbi

Additional

No response

@caarmeecoorbii caarmeecoorbii added the question A HUB question that does not involve a bug label Apr 24, 2024
Copy link

👋 Hello @caarmeecoorbii, thank you for raising an issue about Ultralytics HUB 🚀! Please visit our HUB Docs to learn more:

  • Quickstart. Start training and deploying YOLO models with HUB in seconds.
  • Datasets: Preparing and Uploading. Learn how to prepare and upload your datasets to HUB in YOLO format.
  • Projects: Creating and Managing. Group your models into projects for improved organization.
  • Models: Training and Exporting. Train YOLOv5 and YOLOv8 models on your custom datasets and export them to various formats for deployment.
  • Integrations. Explore different integration options for your trained models, such as TensorFlow, ONNX, OpenVINO, CoreML, and PaddlePaddle.
  • Ultralytics HUB App. Learn about the Ultralytics App for iOS and Android, which allows you to run models directly on your mobile device.
    • iOS. Learn about YOLO CoreML models accelerated on Apple's Neural Engine on iPhones and iPads.
    • Android. Explore TFLite acceleration on mobile devices.
  • Inference API. Understand how to use the Inference API for running your trained models in the cloud to generate predictions.

If this is a 🐛 Bug Report, please provide screenshots and steps to reproduce your problem to help us get started working on a fix.

If this is a ❓ Question, please provide as much information as possible, including dataset, model, environment details etc. so that we might provide the most helpful response.

We try to respond to all issues as promptly as possible. Thank you for your patience!

@pderrenger
Copy link
Member

Hello Carme Corbi,

Great question! Understanding metrics is crucial for evaluating your model's performance. For detailed explanations on metrics like Box Loss, Object Loss, and others, our Ultralytics HUB Docs offer comprehensive insights. These sections will guide you in interpreting the metrics to assess training progress, detect overfitting, and understand if your dataset is being trained effectively.

You can visit the Ultralytics HUB Docs at https://docs.ultralytics.com/hub for more information on these metrics and tips on improving your training outcomes.

Happy training! 🚀

@caarmeecoorbii
Copy link
Author

Thank you so much! I've checked the Ultralytics Hub Docs and I can't find where the metrics are explained, could you provide me with the link to the site where they explain it?

@pderrenger
Copy link
Member

Hello again!

I'm glad to hear you've looked into the Ultralytics HUB Docs! My apologies for any confusion caused earlier. It seems I can't directly provide links other than to our main documentation page at https://docs.ultralytics.com/hub.

A more detailed exploration within the Docs, especially in the sections related to training, evaluation, and tutorials, should give you insights into understanding and leveraging the metrics for your training sessions. If the specifics on Box Loss and Object Loss aren't directly highlighted, browsing through sections covering model evaluation metrics might offer the information indirectly through broader context.

Hope that helps narrow down your search! Happy exploring! 🕵️‍♂️

@caarmeecoorbii
Copy link
Author

Hello again,

Thank you very much. Is it possible to know how Box Loss and Object Loss are calculated?

@pderrenger
Copy link
Member

Hello!

Absolutely, I'd be happy to explain briefly without diving into code!

  • Box Loss measures the accuracy of the predicted bounding boxes against the ground truth boxes. It takes into account the differences in the center coordinates, width, and height of the predicted versus actual boxes.

  • Object Loss evaluates how well the model predicts the presence of an object within a bounding box. It compares the model's confidence scores against the actual presence (or absence) of objects.

Both losses are crucial for fine-tuning the accuracy of object detection models. For the detailed mathematical formulations and how they're integrated into the training process, I'd recommend a deep dive into the official YOLO papers or object detection literature. 📚

Happy modeling!

@caarmeecoorbii
Copy link
Author

Tranks!

I have just finished training the YOLOv8 detector. I trained it for 100 epochs. Is there any way to take the weights up to epoch 25?

@sergiuwaxmann
Copy link
Member

@caarmeecoorbii There is no way to take the weights from epoch 25 but the final weights are the best weights.

@caarmeecoorbii
Copy link
Author

Thanks! I have another question about trackers. Are only BOTSort and ByteTrack trackers available? Where can I find more information about the architecture of these trackers you use?

@pderrenger
Copy link
Member

Hello!

Yes, within the Ultralytics HUB, BOTSort and ByteTrack are the trackers currently available. For detailed insights on the architecture and implementation of these trackers, the original research papers for BOTSort and ByteTrack offer the most comprehensive information. Although I can't link directly to external sources, searching for these papers by their titles in academic databases or preprint servers like arXiv should lead you to their methodologies, architectures, and performance evaluations.

Hope this points you in the right direction! Happy tracking! 🕵️‍♂️

@caarmeecoorbii
Copy link
Author

What are the best parameters for the tracker ByteTrack if I want to detect small objects?
image

@pderrenger
Copy link
Member

Hello!

For detecting small objects with ByteTrack in the Ultralytics HUB, optimizing parameters such as the detection threshold (--conf-thres) and Non-Maximum Suppression (NMS) threshold (--iou-thres) can be particularly effective. Lowering the --conf-thres may help in picking up smaller objects that the model is less confident about, while adjusting the --iou-thres can help in managing how detections are merged.

Experimenting with these parameters should help you tailor ByteTrack's performance to better detect small objects. Happy detecting! 🚀

@sergiuwaxmann sergiuwaxmann added the fixed Bug is resolved label May 7, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
fixed Bug is resolved question A HUB question that does not involve a bug
Projects
None yet
Development

No branches or pull requests

3 participants