Model Evaluation for Computer Vision
26:22
Active Learning in Computer Vision
13:33
Advanced Computer Vision Workflows
11:07
Пікірлер
@fzgarcia
@fzgarcia 5 сағат бұрын
Hello, I have some questions about neural network processing with YOLO... I have a scenario where I need to identify coffee boxes whose only difference is the color to identify one type of coffee as different from another. I have always believed that for performance reasons, network training converts images to grayscale to work on only 1 color channel and does the same when inferring new images. Is this understanding correct? Is it possible to train the models considering the colors, the 3 RGB channels? Or would the best option be to identify the objects (bounding boxes) and perform post-processing with OpenCV, for example, in the bouding box region to identify the closest color? Thanks
@saadashraf1293
@saadashraf1293 21 сағат бұрын
Would this be good for hate meme or offensive meme pictures detection? Memes that have some pictures along with text in them????
@TamikoMccaffree-r5l
@TamikoMccaffree-r5l Күн бұрын
Toy Plains
@ditya3548
@ditya3548 Күн бұрын
insanely goated bro
@aryachowkekar2769
@aryachowkekar2769 Күн бұрын
TypeError: unhashable type: 'numpy.ndarray' (Having this error in custom training process)
@KAZVorpal
@KAZVorpal Күн бұрын
A good question would be whether open AI is ever going to contribute anything real to Ai research, theory, and progress. So far, all they've done is take somebody else's idea - using attention with a pre-trained transformer - and milk it endlessly without any transformative ideas of their own. They have not actually changed the state of the art at all, just re-engineered it and threw ever more resources at it. There is no fundamental difference between GPT 4o and GPT 2. They just keep throwing memory, data, and RAG engineering plug-in gimmicks at it.
@ThomasRobinson-e5x
@ThomasRobinson-e5x Күн бұрын
Taylor Michelle Lopez Lisa Lewis Cynthia
@cagataydemirbas7259
@cagataydemirbas7259 Күн бұрын
Is Annotation Necessary for the Model to Learn the Details in the Image? or beneficial ? for example, when I throw an image and ask, there is a speed sign on the left side, there are children eating ice cream on the right side, the stork is flying above, the weather is sunny, the ground is asphalt and pavement, etc. Do I need to anotate these for the model to learn them?
@KAZVorpal
@KAZVorpal Күн бұрын
Understand that chant GPT can't see or understand the image in any way, at all. What it actually does is pass the image to a completely different system, that that can understand images, and then passes GPT a text description of it. So either you do the annotation, or an RAG plugin has to do the annotation. GPT for cannot see images, make images, hear sounds, speak, remember things, none of it. Those things are all faked using external plugins.
@caliburrito4
@caliburrito4 Күн бұрын
Awesome video. Could you please provide the code to download the final video outputs? Also, how do you overlay the voronoi diagram and the radar style view onto the final output ? I tried to rewrite the code below based on your code from the beginning but the ellipses, labels and triangle don't track the players and the ball correctly. I'm almost there, please advise, thanks! video_info = sv.VideoInfo.from_video_path(SOURCE_VIDEO_PATH) video_sink = sv.VideoSink(TARGET_VIDEO_PATH, video_info=video_info) frame_generator = sv.get_video_frames_generator(SOURCE_VIDEO_PATH) with video_sink: for frame in tqdm(frame_generator, total=video_info.total_frames): result = PLAYER_DETECTION_MODEL.infer(frame, confidence=0.3)[0] detections = sv.Detections.from_inference(result) labels = [ f"#{tracker_id}" for tracker_id in all_detections.tracker_id ] all_detections.class_id = all_detections.class_id.astype(int) annotated_frame = frame.copy() annotated_frame = ellipse_annotator.annotate( scene=annotated_frame, detections=all_detections) annotated_frame = label_annotator.annotate( scene=annotated_frame, detections=all_detections, labels=labels) annotated_frame = triangle_annotator.annotate( scene=annotated_frame, detections=ball_detections, ) video_sink.write_frame(annotated_frame)
@caliburrito4
@caliburrito4 Күн бұрын
Awesome video. Could you please provide the code to download the final video outputs? Also, how do you overlay the voronoi diagram and the radar style view onto the final output ? I tried to rewrite the code below based on your code from the beginning but the ellipses, labels and triangle don't track the players and the ball correctly. I'm almost there, please advise, thanks! video_info = sv.VideoInfo.from_video_path(SOURCE_VIDEO_PATH) video_sink = sv.VideoSink(TARGET_VIDEO_PATH, video_info=video_info) frame_generator = sv.get_video_frames_generator(SOURCE_VIDEO_PATH) with video_sink: for frame in tqdm(frame_generator, total=video_info.total_frames): result = PLAYER_DETECTION_MODEL.infer(frame, confidence=0.3)[0] detections = sv.Detections.from_inference(result) labels = [ f"#{tracker_id}" for tracker_id in all_detections.tracker_id ] all_detections.class_id = all_detections.class_id.astype(int) annotated_frame = frame.copy() annotated_frame = ellipse_annotator.annotate( scene=annotated_frame, detections=all_detections) annotated_frame = label_annotator.annotate( scene=annotated_frame, detections=all_detections, labels=labels) annotated_frame = triangle_annotator.annotate( scene=annotated_frame, detections=ball_detections, ) video_sink.write_frame(annotated_frame)
@JuanCarlos-ff6vp
@JuanCarlos-ff6vp Күн бұрын
Can you download the trained model (weigjts) ? i want to get the .pt file (for pytorch)
@DanetteKshlerin
@DanetteKshlerin Күн бұрын
Jackson Christopher White Elizabeth Walker Brian
@culture2355
@culture2355 Күн бұрын
Tutorial starts at 14:21
@ramkumarkoppu
@ramkumarkoppu Күн бұрын
Next, plan for llama vlm fine tuning with roboflow datasets, probably object detection
@Luca-v1t
@Luca-v1t 2 күн бұрын
Hey thanks for the Tutorial! I tried to follow your steps, but i've got an error at uploading my trained model to roboflow at 11:30. The message is: An error occured when getting the model upload URL: 404 Client Error: Not Found for url: ... Any ideas?
@14types
@14types 2 күн бұрын
how he said - easy raid?
@Roboflow
@Roboflow 2 күн бұрын
I messed up the name. Fast ReID: github.com/JDAI-CV/fast-reid
@manikanthgoud9411
@manikanthgoud9411 2 күн бұрын
So what if I dont have a model, what do I do? could you please help me
@douradesh
@douradesh 2 күн бұрын
FUTBol
@ramkumarkoppu
@ramkumarkoppu 2 күн бұрын
What is the object detection inference latency with this comparable to YOLO-world?
@SkalskiP
@SkalskiP 2 күн бұрын
It's slooooooow. 5-10s in my experience.
@ramkumarkoppu
@ramkumarkoppu 2 күн бұрын
@@SkalskiP probably run yolo in the front end for realtime and cascade with GPT-4o for zero shot?
@hmp01
@hmp01 3 күн бұрын
bro can I use this to make a football game. I am not a developer of any kind, I want a full realistic sim football game
@kunjanparikh5848
@kunjanparikh5848 3 күн бұрын
Hi can anyone tell me how to chose which yaml file is best suitable for our model?
@14types
@14types 3 күн бұрын
I want to put a camera at the front door and identify visitors - my family/stranger. What is the best way to do this?
@Roboflow
@Roboflow 3 күн бұрын
Make some photos. Create dataset (probably 100 images for start). Train a nano model. Deploy raspberry pi.
@14types
@14types 3 күн бұрын
@@Roboflow can yolo11 do this? I've just never seen Yolo distinguish people's faces. Usually it just distinguishes between a person/dog, etc. i work with esp32cam, not raspberry.
@14types
@14types 3 күн бұрын
@@Roboflow I will send the photos to the server and then pass them through the neural network
@umer6500
@umer6500 3 күн бұрын
What's the secret roboflow key
@Roboflow
@Roboflow 3 күн бұрын
YOLO11 notebook contains information on how to retrieve it
@StacyHamilton-c3u
@StacyHamilton-c3u 3 күн бұрын
Bethel Junctions
@viveksapkale1577
@viveksapkale1577 4 күн бұрын
Your video provides valuable insights into computer vision, and we truly appreciate the depth of information you share. it helps many of us learn and grow. Thank you for your dedication and efforts! 😊
@t_o_ny
@t_o_ny 4 күн бұрын
Crushing it Piotr!!!!!
@ramazanoguz-q4n
@ramazanoguz-q4n 4 күн бұрын
I am working on a project and considering using Roboflow. In my project, I will classify the geometric shapes (triangle, quadrilateral, circle, zigzag, etc.) found on shoe soles. At this point, I will label them using object detection. What is your suggestion on this?
@buzzinglikeabee83
@buzzinglikeabee83 5 күн бұрын
ipad is not supported in using cvat 😢
@B4l0ur
@B4l0ur 5 күн бұрын
How can I then deploy the trained model on roboflow pls ? It is said it is not supported and if i canhed to yolov8 there is a depedency issue : ultralytics==8.0.196 is required whereas ultralytics==8.3.2 is installed
@Roboflow
@Roboflow 5 күн бұрын
We are still working on it ;) stay tuned
@mikeazatov1537
@mikeazatov1537 5 күн бұрын
Was something in Roboflow package updated recently? Getting an error trying to download the keypoint dataset. RuntimeError Traceback (most recent call last) <ipython-input-5-0647745e9d8d> in <cell line: 9>() 7 project = rf.workspace("roboflow-jvuqo").project("football-field-detection-f07vi") 8 version = project.version(12) ----> 9 dataset = version.download("yolov8") /usr/local/lib/python3.10/dist-packages/roboflow/core/version.py in download(self, model_format, location, overwrite) 241 else: 242 try: --> 243 raise RuntimeError(response.json()) 244 except json.JSONDecodeError: 245 response.raise_for_status() RuntimeError: {'error': 'Unknown error'}
@ehsan4641
@ehsan4641 5 күн бұрын
Hi, Thanks for your very nice video. I see you are using RTMDET from mmyolo (not from mmdetection). So, I was wondering if it is still free to use it for commercial. I mean the licence of mmyolo is different than mmdetection, and you used mmyolo model in your video. So, has RTMDET from mmyolo the same licence of RTMDET from mmdetection which was "Apache License 2.0"?
@jamzydarth7746
@jamzydarth7746 5 күн бұрын
How would i go about measuring the duration of each class that appears in a zone.
@최일웅-c7w
@최일웅-c7w 5 күн бұрын
At which point do I have to start watching to skip over the training part? I am getting confused because the material is very difficult to understand. Also have you provided all the files and resources to start off from that point?
@felixkuria1250
@felixkuria1250 5 күн бұрын
Can I use Yolo11 for tracking
@MathaGoram
@MathaGoram 5 күн бұрын
As someone who is at the starting gates for YOLOv8 and Roboflow, many, many thanks for publishing these tutorials. You not only keep the presentation at a level that is easily understood (by a dinosaur like me from the punched cards era) but also make it exciting to try out the applications at our own pace owing to superb post-publication support. Best regards.
@37devesh
@37devesh 5 күн бұрын
Breaking News: YOLO v11 Outperforms Previous Versions in Object Detection! Please see Comparision of YOLO v11, YOLO v10, and YOLO v8 across various metrics, including mAP, parameters, and real-world performance. Key Findings: Superior Accuracy: YOLO v11 demonstrated significantly higher accuracy compared to its predecessors on both COCO and OBB-Dota V1 datasets. Check YOLO v11 video tutorial and Colab Demo, where they showcase live demos and discuss the results obtained with YOLO Nano on satellite data. kzbin.info/www/bejne/g52ZnKefotN9gqM #YOLO #deeplearning #objectdetection #yolo_v11 #yolo_v10 #yolo_v8 #computervision #innovation
@AkankshyaParida-z1m
@AkankshyaParida-z1m 5 күн бұрын
i am getting error again and again if i use any other dataset , my epochs are not running only.. suggest what to do
@OnuralpSEZER
@OnuralpSEZER 6 күн бұрын
Thank you so much for stream 🎉😊
@benephr
@benephr 6 күн бұрын
Hi, can I input a video file of a full match and let it generate a video containing all touches by a certain player?
@yeongnamtan
@yeongnamtan 6 күн бұрын
I get below error when trying time-in-zone... ARN:[email protected]] global cap_msmf.cpp:488 `anonymous-namespace'::SourceReaderCB::OnReadSample videoio(MSMF): async ReadSample() call is failed with error status: -1072875772 [ WARN:[email protected]] global cap_msmf.cpp:1769 CvCapture_MSMF::grabFrame videoio(MSMF): can't grab frame. Error: -1072875772
@benephr
@benephr 6 күн бұрын
i lost it when "it's called football"😂THAT'S RIGHT BROOOO!
@PhilemonVic-o5i
@PhilemonVic-o5i 7 күн бұрын
Martinez Larry Martin Edward Clark Kimberly
@robertodavinn
@robertodavinn 7 күн бұрын
Hello, are you able to make an updated guide regarding training custom datasets into the EfficientDet model? It seems like most functions have been deprecated. Thank you in advance.
@RonaldGonzalez-l9h
@RonaldGonzalez-l9h 7 күн бұрын
Meghan Stravenue
@nafiserfan3576
@nafiserfan3576 7 күн бұрын
I encounter an error while trying to continue deploying the trained model with the current version of ultralytics (8.2.103) instead of the said dependency (8.0.196). Suggestions?
@Roboflow
@Roboflow 7 күн бұрын
Hi! It’s just a warning. No need to worry about it. Just confirm.
@nafiserfan3576
@nafiserfan3576 7 күн бұрын
@@Roboflow leads to an error not a warning for me: "404 Client Error: Not Found for url: ..."
@ivancoppola01
@ivancoppola01 3 күн бұрын
@@nafiserfan3576 Same for me
@minditon3264
@minditon3264 8 күн бұрын
Hi Is it possible to get the GitHub repo or other code associated with the working of Florence
@a1mae
@a1mae 8 күн бұрын
our project is to detect and count the object on the captured photo. can we follow this tutorial? or is there other more applicable tutorial we can follow
@GunaNadarWL
@GunaNadarWL 8 күн бұрын
how much will it cost for a single stream for a month with maybe 2 fps provided we have already trained the model with roboflow?
@giancarlocestarolli6049
@giancarlocestarolli6049 9 күн бұрын
Hey man, i'm back here! Can you help me turning the Radar-view and Voronoi Diagram into video ? There's a long time that i'm trying it, but it always generates a 258Bytes video...