guides/nvidia-jetson/ #9942
Replies: 30 comments 66 replies
-
Which one reduce inference time using TensorRT or deepstream? |
Beta Was this translation helpful? Give feedback.
-
Hi, for this docker that run on jetpack 4x, I need to use csi camera for my project, and I face a lot of problems about it. Without docker, I can open the camera. With docker, I cant open it although I already add this command --volume /tmp/argus_socket:/tmp/argus_socket. This is my command to activate ultralytics docker: I not sure is the torch and torchvision version that beyond the Jetpack 4 which is 1.11.0 that suppose for Jetpack 5 affect the csi camera. Please help me, thanks. |
Beta Was this translation helpful? Give feedback.
-
Hi, I am using jetson nano with jetpack4 and jetson orin nano with jetpack5.1.3. I am trying to use webcamera inside the docker and with the follow argument it works for jetson nano jetpack4--> sudo docker run -it --runtime nvidia My question is: How it change for jetson orin nano? how can I change the last line? " ultralytics/ultralytics:latest-jetson-jetpack4 Thanks in advanced |
Beta Was this translation helpful? Give feedback.
-
Good day, as i can see in one of the above pictures, there's 15G of memory available which is quite alright. PS: I intend to detect in real time for hours. |
Beta Was this translation helpful? Give feedback.
-
I tried ultralytics for jetson orin nano "Start without Docker" and I am getting the follow error: (env3.8) jetson-orin@jetsonorin-desktop:~/Desktop/gpu$ python3 Traceback (most recent call last): I am seeing is not only my problem and there is a post in NVIDIA forum: https://forums.developer.nvidia.com/t/pytorch-installation-on-jetson-agx-orin/292846 help: last line of installation: python3 setup.py install --user |
Beta Was this translation helpful? Give feedback.
-
Hello! I'm running YOLOv8 on my Jetson Nano JetPack 4.x. I took the Docker approach.
These results are unexpected. You'd think TensorRT/.engine would be faster but in fact it's much slower. I'm wondering if this should raise any red flags? And if so, how might I proceed? I also ask because I've been having a lot of trouble exporting YOLOv8 on the Nano so just wanted to be on the lookout for any potential problems so that I can catch them early on. Thanks in advance! |
Beta Was this translation helpful? Give feedback.
-
when i run this command on Jetson Xavier NX (Jetpack 5.1.2) What should i do? |
Beta Was this translation helpful? Give feedback.
-
WARNING I am getting this warning when i ran these code Load a YOLOv8n PyTorch modelmodel = YOLO("yolov8n.pt") Export the modelmodel.export(format="engine") " Even though i have my torch cuda enabled. |
Beta Was this translation helpful? Give feedback.
-
Is there any intention to update the documentation for Jetpack 6? |
Beta Was this translation helpful? Give feedback.
-
Hi, I plan on using the docker that'll run on jetpack 4.x and would need to use a CSI camera (IMX519) for my project. As it is my first time using a docker, are there any videos that will demonstrate a step process on using a docker on a Jetson Nano? |
Beta Was this translation helpful? Give feedback.
-
hi , i have error: torchvision 0.16.2 requires torch==2.1.2, but you have torch 2.1.0a0+41361538.nv23.6 which is incompatible. |
Beta Was this translation helpful? Give feedback.
-
Yes, I follow this but have this error
Vào Th 6, 2 thg 8, 2024 lúc 12:11 Glenn Jocher ***@***.***>
đã viết:
… @Faker-T1-NO1 <https://github.com/Faker-T1-NO1> to resolve the
compatibility issue between torchvision and torch, you need to ensure
both packages are compatible with each other. Uninstall the current
versions and install the compatible versions as follows:
1.
Uninstall the existing packages:
pip uninstall torch torchvision
2.
Install the compatible versions:
sudo apt-get install -y libopenblas-base libopenmpi-dev
wget https://developer.download.nvidia.com/compute/redist/jp/v512/pytorch/torch-2.1.0a0+41361538.nv23.06-cp38-cp38-linux_aarch64.whl
pip install torch-2.1.0a0+41361538.nv23.06-cp38-cp38-linux_aarch64.whl
sudo apt install -y libjpeg-dev zlib1g-dev
git clone https://github.com/pytorch/vision torchvisioncd torchvision
git checkout v0.16.2
python3 setup.py install --user
This should resolve the version mismatch. For more details, refer to the NVIDIA
Jetson setup guide <https://docs.ultralytics.com/guides/nvidia-jetson/>.
—
Reply to this email directly, view it on GitHub
<#9942 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/BKIS5BJS3DHXAF74PQRH5FDZPMIINAVCNFSM6AAAAABGAK2IGGVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTAMRRHE2DQNQ>
.
You are receiving this because you were mentioned.Message ID:
***@***.***
com>
|
Beta Was this translation helpful? Give feedback.
-
When i try to run "results = model.benchmarks(data="coco8.yaml", imgsz=640)" Traceback (most recent call last): How do i resolve it, I am doing it on Jetson Xavier NX? |
Beta Was this translation helpful? Give feedback.
-
Hello , for docker version and jetpack4x on jetson nano this is my command to run
I am not getting a display output even though the code is running and detection results r printed on the terminal I assume that
the above line is the one to support display and i have included it How to fix that what changes should i make to the command OR any other changes needed |
Beta Was this translation helpful? Give feedback.
-
Hello, PyTorch: starting from 'yolov8n.pt' with input shape (1, 3, 640, 640) BCHW and output shape(s) (1, 84, 8400) (6.2 MB) ONNX: starting export with onnx 1.16.2 opset 17... ONNX: export success ✅ 1.4s, saved as 'yolov8n.onnx' (12.2 MB) × python setup.py egg_info did not run successfully. note: This error originates from a subprocess, and is likely not a problem with pip. × Encountered error while generating package metadata. note: This is an issue with the package mentioned above, not pip. × python setup.py egg_info did not run successfully. note: This error originates from a subprocess, and is likely not a problem with pip. × Encountered error while generating package metadata. note: This is an issue with the package mentioned above, not pip. Please tell me if you need any more information. I'd really appreciate some help |
Beta Was this translation helpful? Give feedback.
-
I would like to ask why I deployed my self trained m model on Jetson Orin NX, and it takes 50ms for frame inference and 30ms for frame inference in FP16 |
Beta Was this translation helpful? Give feedback.
-
My device is a jetson nano, can I use the gpu version of pytorch after I pull the docker image? |
Beta Was this translation helpful? Give feedback.
-
Hello ! I am using YOLOv8 with Docker (*) on Jetson Nano Jetpack 4.6.1 [L4T 32.7.1] I had problem with Python to display the results. Here is an example: from ultralytics import YOLO Load a modelmodel = YOLO("yolov8n.pt") # pretrained YOLOv8n model Run batched inference on a list of imagesresults = model(["image1.jpg", "image2.jpg"]) # return a list of Results objects Process results listfor result in results: I get: 0: 640x480 4 persons, 1 bus, 1 stop sign, 258.6ms but nothing is displayed. |
Beta Was this translation helpful? Give feedback.
-
How do I use the Jetson.GPIO in a jetpack4 docker image |
Beta Was this translation helpful? Give feedback.
-
If I work on Sahi algorithm with Yolo 8, will these things apply to it? NVIDIA A100: FP32 Inference: ~0.52 ms / image FP32 Inference: ~1.06 ms / image |
Beta Was this translation helpful? Give feedback.
-
If I work on Sahi algorithm with Yolo 8, will these things apply to it? NVIDIA A100: FP32 Inference: ~0.52 ms / image FP32 Inference: ~1.06 ms / image |
Beta Was this translation helpful? Give feedback.
-
How can I deploy YOLO11 model on Jetson TX2? |
Beta Was this translation helpful? Give feedback.
-
Hi, Jetpack 6.1 has been released, the yolo model cannot be trained on this platform, which gives an error.
I assume it is because my cudnn version is updated to 9.3.0 along with Jetpack 6.1. |
Beta Was this translation helpful? Give feedback.
-
I am currently working with the NVIDIA Jetson TX2, which has 8 GB of LPDDR4 memory and 32 GB of eMMC storage. My device runs JetPack 4, and I am deploying YOLOv8 on it. While setting up, I encountered a "disk space 0" error when attempting to run Docker commands, which I suspect is due to storage constraints. Could you advise on the recommended free space required for running a pre-built Docker image of JetPack 4 with YOLOv8? Additionally, I have configured YOLOv8 in a CUDA-enabled virtual environment. However, I am experiencing issues linking TensorRT to this environment, which is preventing me from converting the YOLOv8 Could you recommend solutions or steps for successfully performing this model conversion, either within the virtual environment or using Docker? |
Beta Was this translation helpful? Give feedback.
-
Is there a way to verify how many YOLOv8n/YOLO11n models the Jetson AGX Orin 64G can run simultaneously, or does it require testing by incrementally adding models one at a time? |
Beta Was this translation helpful? Give feedback.
-
Hello!! ... I want to set up Ultralytics and PyTorch in JetPack 6.1 on a Jetson AGX Orin. I followed your guide, but it’s not working well.
But here’s where the issue starts! pip’s dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. After installing torch, torchvision and trying to run Ultralytics, I get this same error again. ImportError: libcudnn.so.8: cannot open shared object file: No such file or directory |
Beta Was this translation helpful? Give feedback.
-
Hi. I wonder if the model exported using tensorRT is the same as the normal model object. For example, model = YOLO("yolo11n.pt") Export the model to TensorRT with DLA enabled (only works with FP16 or INT8)model.export(format="engine", device="dla:0", half=True) # dla:0 or dla:1 corresponds to the DLA cores Load the exported TensorRT modeltrt_model = YOLO("yolo11n.engine") Run inferenceresults = trt_model("https://ultralytics.com/images/bus.jpg") I wonder if trt_model in the code above is ultralytics.engine.model.model class. If it is the same, is it correct to use methods such as load, save, predict, and train in this model object? If not, please let me know the page with specific information about this. Thanks! |
Beta Was this translation helpful? Give feedback.
-
Hi is the guide updated for Jetpack 6.1 ? it feels like torch version are for jetpack 6.0 (please remove tested for jetpack 6.x) since the new jepack 6.1 comes with CUDA 12.6. where as the torch given on the guide is for 12.2 |
Beta Was this translation helpful? Give feedback.
-
Hey there!
Since Torch seems to be installed successfully, I try to install Torchvision v0.16.2 with "python3 setup.py install --user" but an new issue:
Then I have tried to install with "pip install ." according this link.
Some information:
Can you help me? Thanks! |
Beta Was this translation helpful? Give feedback.
-
jetson nx information:
Platform
Machine: aarch64
System: Linux Hardware
Model: NVIDIA Orin NX Developer
KitDistribution:Ubuntu 20.04 focal 699-level Part Number:699-13767-0000-300 P.1
Release:5.10.104-tegra P-Number: p3767-0800
Python: 3.9.20 Module:NVIDIA Jetson Orin NX(16GB ram)
Soc: tegra23x
CUDA Arch BIN: 8.7
Libraries Codename:P3768
CUDA: 11.4.315 L4T: 35.3.1
CuDNN:8.6.0.166 Jetpack:5.1.1
TensorRT:8.5.2.2
VPI:2.2.7
Vulkan:1.3.204
Opency:4.5.4 with CUDA: NO
docker pull
ultralytics/ultralytics:latest-jetson-jetpack5
sudo docker pull ultralytics/ultralytics:latest-jetson-jetpack5
sudo docker run -it --ipc=host --runtime=nvidia ultralytics/ultralytics:latest-jetson-jetpack5
root@ubuntu:/ultralytics# yolo export model=yolo11n.pt format=engine
WARNING ⚠️ TensorRT requires GPU export, automatically assigning device=0
Ultralytics 8.3.51 🚀 Python-3.8.10 torch-2.4.1
Traceback (most recent call last):
File "/usr/local/bin/yolo", line 8, in <module>
sys.exit(entrypoint())
File "/ultralytics/ultralytics/cfg/__init__.py", line 972, in entrypoint
getattr(model, mode)(**overrides) # default args from model
File "/ultralytics/ultralytics/engine/model.py", line 738, in export
return Exporter(overrides=args, _callbacks=self.callbacks)(model=self.model)
File "/ultralytics/ultralytics/engine/exporter.py", line 225, in __call__
self.device = select_device("cpu" if self.args.device is None else self.args.device)
File "/ultralytics/ultralytics/utils/torch_utils.py", line 192, in select_device
raise ValueError(
ValueError: Invalid CUDA 'device=0' requested. Use 'device=cpu' or pass valid CUDA device(s) if available, i.e. 'device=0' or 'device=0,1,2,3' for Multi-GPU.
torch.cuda.is_available(): False
torch.cuda.device_count(): 0
os.environ['CUDA_VISIBLE_DEVICES']: None
See https://pytorch.org/get-started/locally/ for up-to-date torch install instructions if no CUDA devices are seen by torch.
Is there something wrong with my docker image version? Not compatible with my device? |
Beta Was this translation helpful? Give feedback.
-
guides/nvidia-jetson/
Quick start guide to setting up YOLOv8 on a NVIDIA Jetson device with comprehensive benchmarks.
https://docs.ultralytics.com/guides/nvidia-jetson/
Beta Was this translation helpful? Give feedback.
All reactions