tasks/detect/ #8489
Replies: 60 comments 202 replies
-
Will color differences in similar objects have an effect? For example, I created a model using black horse data and I will use it for white horses detection |
Beta Was this translation helpful? Give feedback.
-
I've finished trained my v8 detection model and i wanted to used a 3d camera (intel Realsense d435) to do inference on live data and then find the XYZ of the the centroid of the object boundary box in reference to the camera origin. How would i do this? |
Beta Was this translation helpful? Give feedback.
-
So is there a way to 'lock on' one particular detected object? Say I have the yolov8 detected 10 objects, my program then go through these objects and determine one particular object I want to do a 'lock' on it. Is there a way I can use yolov8 from that point to ignore the rest of the detected objects and just keep tracking/locking on the particular object I want? Thanks in advance. |
Beta Was this translation helpful? Give feedback.
-
It looks like I can not tell Yolov8 to stop other detections and just focus/locking on a particular detected object in real-time. Yes, it seems that a different tracking/locking algorithm needs to be in place for locking before I am done with it. I sure wish Yolov could have this ability to do so, I guess I can put this in a wish list.... Thanks for your reply. |
Beta Was this translation helpful? Give feedback.
-
Yes, thank you for your suggestion. I realized the tracking option in Yolov8 is there but still, it tracks multiple instead of just one. My program still needs filter logic to catch what I wanted. I am not sure how Yolov internal works but I presume excluding object by tracking ID will be much more faster than my python filtering logic...anyway, there will be a ton of applications can use the locking facility if provided, I sincerely wish you all can put this into you upcoming enhancements. Thanks for the reply. |
Beta Was this translation helpful? Give feedback.
-
Hey I am very much confused with the overall working of the YOLO model . I mean like when we say each cell is responsible for predicting an image what it actually means in terms of workflow and like what is the concept of anchor boxes like what do you mean by predefined anchor boxes. What i understand is that we have a image like (480,480,3) we finally convert it to like (7,7,30) by passing through convolutional layers / activation layers / pooiling layers . Assuming we divided it into 7 * 7 grid and for each cell we have a structure like lets say one anchor box x,y,w,h,conf,classess) but it is all prediction right ? we didnt defined anything so what does predefined anchor boxes mean? Also want to know one thing how are we suppossed to do transfer learning in object detection i dont get it . But how do we use transfer-learning in object detection i still don get it ? how is it divided into different stages ? Also in this doc for training yolov8 on custom dataset we are training all the layers which isnt transfer learning right and training all layers on a very small dataset wouldnt make sense right ? Can you please please please clear my confusion and share some updated/latest resources which could help me with this . I really need help in this. |
Beta Was this translation helpful? Give feedback.
-
Hi, for det in results[0].boxes.xyxy:
xmin, ymin, xmax, ymax, conf = map(int, det[:4].tolist() + [det[4]]) xyxy in 1,4 dimensional and therefor det[4] is out of bounds. what i did manage to get working but it feels convoluted since I am using [0] so many times for such a short code is from ultralytics import YOLO
import cv2
model = YOLO('./runs/detect/train/weights/best.pt')
result = model('tennis.jpg')
for result in result[0].boxes.numpy():
xyxy = result.xyxy[0]
x1, y1, x2, y2 = map(float, xyxy)
cls = int(result.cls[0])
conf = float(result.conf[0])
print("x1: ", x1, "y1: ", y1, "x2: ", x2, "y2: ", y2, "cls: ", cls, "conf: ", conf) |
Beta Was this translation helpful? Give feedback.
-
hi, im implementing yolov8 in my android gallery's search algorithm. the yolov8's purpose is to scan media files and return images that has a box label that matches the search query. i can make it work using yolov5s.torchscript.ptl with org.pytorch:pytorch_android_lite:1.10.0 and org.pytorch:pytorch_android_torchvision_lite:1.10.0 but i cant make it work with yolov8s.torchscript |
Beta Was this translation helpful? Give feedback.
-
Can you provide me with the best configurations for human-detection example using YOLOv8? |
Beta Was this translation helpful? Give feedback.
-
how to reduce predicted bounding box line width in yolov8 |
Beta Was this translation helpful? Give feedback.
-
Hi, can yolov8 detect color sequence?. i create color sequence of rainbow (red, yellow, green) as a real rainbow. if i change the yellow to orange, can yolov8 detect it as a wrong rainbow? |
Beta Was this translation helpful? Give feedback.
-
i need a object detection code using yolov8 for detect the two images |
Beta Was this translation helpful? Give feedback.
-
here is my code import cv2 Load the YOLO modelmodel = YOLO('yolov8n.pt') # Assuming 'yolov8n.pt' is the YOLO model you want to use Specify the path to your imageimage_path = '/media/XavierSSD250/22_08_23/ui.png' Perform object detectionresults = model(image_path) Convert image to OpenCV formatimage = cv2.imread(image_path) Extract bounding boxes, labels, and confidence scores from resultsbboxes = results.pred[0][:, :4] # Extract bounding boxes Draw bounding boxes and labels on the imagefor bbox, label, confidence in zip(bboxes, labels, confidences): Show the image with bounding boxescv2.imshow('Object Detection', image) |
Beta Was this translation helpful? Give feedback.
-
I am a beginner and have trained an object detection model using my own dataset. If I want to display the results of detection inference in real-time on the web, which method would be better if you recommend? |
Beta Was this translation helpful? Give feedback.
-
Hi, at the moment I'm trying to used .onnx exported from yolov8n.pt in Unity, using C#, to demonstrate real-time object detection in 3D scene, I got output in the shape of (1,84,8400). I would like to ask how to post process this output to draw bounding box, extracting the information of the output or the output format. Sorry for my english, and if there are any better method please suggest right away. thank you in advance. |
Beta Was this translation helpful? Give feedback.
-
hello all I have around original images around 74 images dataset with resolution 8192 * 5460 Train the modelmodel.train( below is my results but the model is not performing well, in terms of precision. please give me hints how to imporve the model and how to make use of all the resources to reduce the training model also I have access to powerful machine with 5 GPU but I am not able to utilize all the resources the machine spec: NVIDIA-SMI 525.147.05 Driver Version: 525.147.05 CUDA Version: 12.0 | |
Beta Was this translation helpful? Give feedback.
-
How to disable print outputs like this: “0: 384x640 3 bags, 124.8ms |
Beta Was this translation helpful? Give feedback.
-
Hello, I want to improve the detection accuracy of specific classes in the dataset without using retraining. My current idea is to increase the weight of specific classes in the loss function. Is this practice effective? Are there other effective ways to do this? |
Beta Was this translation helpful? Give feedback.
-
Hi I have a project where we intend to place cameras in warehouses. take periodic snapshots and find differences in images to see what has entered the warehouse. can yolo do this? |
Beta Was this translation helpful? Give feedback.
-
How can I use yolo V8 with amd Radeon r5 m335 |
Beta Was this translation helpful? Give feedback.
-
I have trained the YOLO11l model for 1000 epochs, and now I want to continue training for 1500 epochs. I passed resume=True, but it is not working. I used a checkpoint with the saved optimizer state. However, with the YOLOv8 model, I am able to continue the training for additional epochs without any issues. -> model = YOLO('checkpoint/epoch998.pt') |
Beta Was this translation helpful? Give feedback.
-
I have trained YOLOv8 on my custom dataset, and it’s working well. However, I’m facing a challenge: I’ve captured a video where the objects are in fixed positions, but the camera is moving. My goal is to get the total count of all detected objects. The problem is that I’m getting the object count for every frame, but I need the actual overall count of the objects, not repeated detections from each frame. How can I resolve this? |
Beta Was this translation helpful? Give feedback.
-
If i use the scale parameter while training my model, will this scale every image in the dataset? or will it train itself on the regular images and then go back through and train itself again with all images scaled? Also is the only difference between YOLO11 and YOLO8 how accurate they are or fast they are? we are currently using YOLOv8 for object detection but if YOLO11 is faster and more accurate we will switch before we train. Thank you! |
Beta Was this translation helpful? Give feedback.
-
Hello, good morning. After doing my detection training in yolov8, I ran my training set images and saved the predicted images with the bounding boxes, saved the txt file with the labels returned from the bounding boxes (returns the ) and saved the file with the boxes that returns all the data for xyxy, xyxyn, xywh, xywhn in a more complete way. After doing this, I want to calculate if there is overlap of the bounding boxes, I thought about doing it by analyzing if the point of one box is inside another. Do you have any other suggestions on how to obtain this result of the overlap of the boxes? And how can I read the file and obtain the necessary data? |
Beta Was this translation helpful? Give feedback.
-
hello, I trained yolo11 with 27 class, and raw output[0] for detection is 51x31x8400, how I do get final bbox, cls and conf from this? I assume there are 51 detection heads, 4 bbox coordinates + 27 class scores, and 8400 anchor points. |
Beta Was this translation helpful? Give feedback.
-
Hello Thanks you |
Beta Was this translation helpful? Give feedback.
-
Saved onnx format model is not working as expected. Getting output shape (1,10,8400). How to fix it to get boxs, and classes.
|
Beta Was this translation helpful? Give feedback.
-
Within my tracking file I have the following code, results = detection.predict(video_path, stream=True, save=False, max_det=2, verbose=False, save_crop=True, project=output, agnostic_nms=True, iou=0.5) This snippet of code is used to calculate the distances of the the two objects in each frame of the output. Our model however is having some trouble distinguishing the two objects (they are very very similar and have the same background), for example it will correctly detect that there are two objects within that frame but will give them the same object label. I was thinking that within my |
Beta Was this translation helpful? Give feedback.
-
Can we use tracker mode other than bot sort and bytetrack? for example with deepsort tracker. |
Beta Was this translation helpful? Give feedback.
-
tasks/detect/
Official documentation for YOLOv8 by Ultralytics. Learn how to train, validate, predict and export models in various formats. Including detailed performance stats.
https://docs.ultralytics.com/tasks/detect/
Beta Was this translation helpful? Give feedback.
All reactions