tasks/detect/ #8489
Replies: 70 comments 236 replies
-
|
Will color differences in similar objects have an effect? For example, I created a model using black horse data and I will use it for white horses detection |
Beta Was this translation helpful? Give feedback.
-
|
I've finished trained my v8 detection model and i wanted to used a 3d camera (intel Realsense d435) to do inference on live data and then find the XYZ of the the centroid of the object boundary box in reference to the camera origin. How would i do this? |
Beta Was this translation helpful? Give feedback.
-
|
So is there a way to 'lock on' one particular detected object? Say I have the yolov8 detected 10 objects, my program then go through these objects and determine one particular object I want to do a 'lock' on it. Is there a way I can use yolov8 from that point to ignore the rest of the detected objects and just keep tracking/locking on the particular object I want? Thanks in advance. |
Beta Was this translation helpful? Give feedback.
-
|
It looks like I can not tell Yolov8 to stop other detections and just focus/locking on a particular detected object in real-time. Yes, it seems that a different tracking/locking algorithm needs to be in place for locking before I am done with it. I sure wish Yolov could have this ability to do so, I guess I can put this in a wish list.... Thanks for your reply. |
Beta Was this translation helpful? Give feedback.
-
|
Yes, thank you for your suggestion. I realized the tracking option in Yolov8 is there but still, it tracks multiple instead of just one. My program still needs filter logic to catch what I wanted. I am not sure how Yolov internal works but I presume excluding object by tracking ID will be much more faster than my python filtering logic...anyway, there will be a ton of applications can use the locking facility if provided, I sincerely wish you all can put this into you upcoming enhancements. Thanks for the reply. |
Beta Was this translation helpful? Give feedback.
-
|
Hey I am very much confused with the overall working of the YOLO model . I mean like when we say each cell is responsible for predicting an image what it actually means in terms of workflow and like what is the concept of anchor boxes like what do you mean by predefined anchor boxes. What i understand is that we have a image like (480,480,3) we finally convert it to like (7,7,30) by passing through convolutional layers / activation layers / pooiling layers . Assuming we divided it into 7 * 7 grid and for each cell we have a structure like lets say one anchor box x,y,w,h,conf,classess) but it is all prediction right ? we didnt defined anything so what does predefined anchor boxes mean? Also want to know one thing how are we suppossed to do transfer learning in object detection i dont get it . But how do we use transfer-learning in object detection i still don get it ? how is it divided into different stages ? Also in this doc for training yolov8 on custom dataset we are training all the layers which isnt transfer learning right and training all layers on a very small dataset wouldnt make sense right ? Can you please please please clear my confusion and share some updated/latest resources which could help me with this . I really need help in this. |
Beta Was this translation helpful? Give feedback.
-
|
Hi, for det in results[0].boxes.xyxy:
xmin, ymin, xmax, ymax, conf = map(int, det[:4].tolist() + [det[4]])xyxy in 1,4 dimensional and therefor det[4] is out of bounds. what i did manage to get working but it feels convoluted since I am using [0] so many times for such a short code is from ultralytics import YOLO
import cv2
model = YOLO('./runs/detect/train/weights/best.pt')
result = model('tennis.jpg')
for result in result[0].boxes.numpy():
xyxy = result.xyxy[0]
x1, y1, x2, y2 = map(float, xyxy)
cls = int(result.cls[0])
conf = float(result.conf[0])
print("x1: ", x1, "y1: ", y1, "x2: ", x2, "y2: ", y2, "cls: ", cls, "conf: ", conf) |
Beta Was this translation helpful? Give feedback.
-
|
hi, im implementing yolov8 in my android gallery's search algorithm. the yolov8's purpose is to scan media files and return images that has a box label that matches the search query. i can make it work using yolov5s.torchscript.ptl with org.pytorch:pytorch_android_lite:1.10.0 and org.pytorch:pytorch_android_torchvision_lite:1.10.0 but i cant make it work with yolov8s.torchscript |
Beta Was this translation helpful? Give feedback.
-
|
Can you provide me with the best configurations for human-detection example using YOLOv8? |
Beta Was this translation helpful? Give feedback.
-
|
how to reduce predicted bounding box line width in yolov8 |
Beta Was this translation helpful? Give feedback.
-
|
Hi, can yolov8 detect color sequence?. i create color sequence of rainbow (red, yellow, green) as a real rainbow. if i change the yellow to orange, can yolov8 detect it as a wrong rainbow? |
Beta Was this translation helpful? Give feedback.
-
|
i need a object detection code using yolov8 for detect the two images |
Beta Was this translation helpful? Give feedback.
-
|
here is my code import cv2 Load the YOLO modelmodel = YOLO('yolov8n.pt') # Assuming 'yolov8n.pt' is the YOLO model you want to use Specify the path to your imageimage_path = '/media/XavierSSD250/22_08_23/ui.png' Perform object detectionresults = model(image_path) Convert image to OpenCV formatimage = cv2.imread(image_path) Extract bounding boxes, labels, and confidence scores from resultsbboxes = results.pred[0][:, :4] # Extract bounding boxes Draw bounding boxes and labels on the imagefor bbox, label, confidence in zip(bboxes, labels, confidences): Show the image with bounding boxescv2.imshow('Object Detection', image) |
Beta Was this translation helpful? Give feedback.
-
|
I am a beginner and have trained an object detection model using my own dataset. If I want to display the results of detection inference in real-time on the web, which method would be better if you recommend? |
Beta Was this translation helpful? Give feedback.
-
|
Hi, at the moment I'm trying to used .onnx exported from yolov8n.pt in Unity, using C#, to demonstrate real-time object detection in 3D scene, I got output in the shape of (1,84,8400). I would like to ask how to post process this output to draw bounding box, extracting the information of the output or the output format. Sorry for my english, and if there are any better method please suggest right away. thank you in advance. |
Beta Was this translation helpful? Give feedback.
-
|
Hello Thanks you |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
|
Saved onnx format model is not working as expected. Getting output shape (1,10,8400). How to fix it to get boxs, and classes. |
Beta Was this translation helpful? Give feedback.
-
|
Can i use detect task to train sign language model? I'm using BISINDO sign language which uses two hand to represent an alphabet |
Beta Was this translation helpful? Give feedback.
-
|
Within my tracking file I have the following code, results = detection.predict(video_path, stream=True, save=False, max_det=2, verbose=False, save_crop=True, project=output, agnostic_nms=True, iou=0.5) This snippet of code is used to calculate the distances of the the two objects in each frame of the output. Our model however is having some trouble distinguishing the two objects (they are very very similar and have the same background), for example it will correctly detect that there are two objects within that frame but will give them the same object label. I was thinking that within my |
Beta Was this translation helpful? Give feedback.
-
|
Can we use tracker mode other than bot sort and bytetrack? for example with deepsort tracker. |
Beta Was this translation helpful? Give feedback.
-
|
Is there anyway to ensure that I do not have two bounding boxes appear with the same label within a frame of a video? For example I have two very similar objects and 2 separate labels however on occasion it will put a bounding box around each object with the same label which will skew my data. thank you! |
Beta Was this translation helpful? Give feedback.
-
|
Good day all, trust you are doing great. |
Beta Was this translation helpful? Give feedback.
-
|
Hi!I am using YOLO11 to detect communication signals in their spectrograms. However, incomplete detection often occurs, especially for longer and narrower signals. For example, if the actual signal frame width and height are 1 * 0.1, the detection result is 0.8 * 0.1 or 1 * 0.08. How can I improve the detect results? Or how to avoid incomplete detection in the general field of object detection? I will appreciate your reply! |
Beta Was this translation helpful? Give feedback.
-
|
Hello, Would you happen to know if this information is actually mentioned somewhere here? My question is whether we should annotate an object when it appears partially in the image. |
Beta Was this translation helpful? Give feedback.
-
|
Hi! from ultralytics import YOLO Load a modelmodel = YOLO("yolo11x.pt") # load a pretrained model (recommended for training) def init_distributed_mode(): def cleanup_distributed_mode(): --- Initialize distributed training mode (Run this only ONCE) ---#init_distributed_mode() --- Your training code goes here ---Before retraining or exiting the script, clean upcleanup_distributed_mode() Train the modelresults = model.train(data=".../dataset.yaml", can you tell me by default how many layers of the network are frozen if any? Or all milions of parameters are changin at each epoch?? |
Beta Was this translation helpful? Give feedback.
-
|
Hello, I have a question. How do I get metrics of my pretrained model on test dataset? |
Beta Was this translation helpful? Give feedback.
-
|
Hey guys, I am on a macbook which does not allow us to use Cuda anymore, I know that mac has something called video_toolbox that for me is helping me with acceleration on ffmpeg. What are my otpions to increase the detection speed of my model? Will video_toolbox work within YOLO as well? if so how do i select my model to use that to work faster? |
Beta Was this translation helpful? Give feedback.
-
|
I just trained a yolo11 object detection model on a custom dataset and now I have new images classes to train so how can I train these images without effecting trained weights and if I want to train a class again which I already trained in my model how can I do that basically I want to freeze head,neck and backbone then train so that old parameters stays same and add new weights |
Beta Was this translation helpful? Give feedback.
-
|
I want to detect only 'Person' class. Do I need to modify some yaml file and retrain the model for that - I want to reduce the parameters. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
tasks/detect/
Official documentation for YOLOv8 by Ultralytics. Learn how to train, validate, predict and export models in various formats. Including detailed performance stats.
https://docs.ultralytics.com/tasks/detect/
Beta Was this translation helpful? Give feedback.
All reactions