[{"title":"( 51 个子文件 4.11MB ) pyonnx-example:使用python实现基于onnxruntime的一些模型推断-源码","children":[{"title":"pyonnx-example-master","children":[{"title":"main_detector_.py <span style='color:#111;'> 3.71KB </span>","children":null,"spread":false},{"title":"combine.py <span style='color:#111;'> 7.86KB </span>","children":null,"spread":false},{"title":"collect_hands.py <span style='color:#111;'> 1.67KB </span>","children":null,"spread":false},{"title":"tools","children":[{"title":"show_onnx_graph.py <span style='color:#111;'> 350B </span>","children":null,"spread":false},{"title":"pose_transforms.py <span style='color:#111;'> 2.40KB </span>","children":null,"spread":false},{"title":"remove_initializer_from_input.py <span style='color:#111;'> 904B </span>","children":null,"spread":false},{"title":"visualize.py <span style='color:#111;'> 5.48KB </span>","children":null,"spread":false},{"title":"get_points_from_cocowb.py <span style='color:#111;'> 14.98KB </span>","children":null,"spread":false},{"title":"test_ipcamera.py <span style='color:#111;'> 936B </span>","children":null,"spread":false}],"spread":true},{"title":"inference","children":[{"title":"keypoints_lpn.py <span style='color:#111;'> 2.26KB </span>","children":null,"spread":false},{"title":"base.py <span style='color:#111;'> 2.61KB </span>","children":null,"spread":false},{"title":"keypoints.py <span style='color:#111;'> 6.13KB </span>","children":null,"spread":false},{"title":"detector.py <span style='color:#111;'> 4.23KB </span>","children":null,"spread":false},{"title":"classifier.py <span style='color:#111;'> 1.95KB </span>","children":null,"spread":false},{"title":"__init__.py <span style='color:#111;'> 200B </span>","children":null,"spread":false},{"title":"keypoints_darkpose.py <span style='color:#111;'> 2.17KB </span>","children":null,"spread":false},{"title":"detector_yolov5.py <span style='color:#111;'> 2.10KB </span>","children":null,"spread":false}],"spread":true},{"title":"README.md <span style='color:#111;'> 1.61KB </span>","children":null,"spread":false},{"title":"main_gesture_.py <span style='color:#111;'> 2.11KB </span>","children":null,"spread":false},{"title":"data","children":[{"title":"main_result","children":[{"title":"gesture","children":[{"title":"003.jpg <span style='color:#111;'> 23.34KB </span>","children":null,"spread":false}],"spread":true},{"title":"keypoints","children":[{"title":"close.jpg <span style='color:#111;'> 14.47KB </span>","children":null,"spread":false},{"title":"open.jpg <span style='color:#111;'> 7.11KB </span>","children":null,"spread":false},{"title":"002.jpg <span style='color:#111;'> 37.64KB </span>","children":null,"spread":false},{"title":"004.jpg <span style='color:#111;'> 36.83KB </span>","children":null,"spread":false},{"title":"003.jpg <span style='color:#111;'> 25.00KB </span>","children":null,"spread":false}],"spread":true},{"title":"pose2","children":[{"title":"17demo.gif <span style='color:#111;'> 1.66MB </span>","children":null,"spread":false}],"spread":true},{"title":"classifier","children":[{"title":"open","children":[{"title":"open.jpg <span style='color:#111;'> 6.66KB </span>","children":null,"spread":false}],"spread":false},{"title":"close","children":[{"title":"close.jpg <span style='color:#111;'> 13.50KB </span>","children":null,"spread":false}],"spread":false}],"spread":true},{"title":"detection","children":[{"title":"crop_res","children":[{"title":"tie","children":[{"title":"002.jpg <span style='color:#111;'> 9.33KB </span>","children":null,"spread":false},{"title":"003.jpg <span style='color:#111;'> 3.55KB </span>","children":null,"spread":false}],"spread":false},{"title":"person","children":[{"title":"001.jpg <span style='color:#111;'> 116.73KB </span>","children":null,"spread":false},{"title":"000.jpg <span style='color:#111;'> 78.88KB </span>","children":null,"spread":false}],"spread":false}],"spread":false},{"title":"zidane.jpg <span style='color:#111;'> 236.62KB </span>","children":null,"spread":false},{"title":"004.jpg <span style='color:#111;'> 36.60KB </span>","children":null,"spread":false}],"spread":false},{"title":"pose","children":[{"title":"133demo.gif <span style='color:#111;'> 1.45MB </span>","children":null,"spread":false},{"title":"002.jpg <span style='color:#111;'> 49.05KB </span>","children":null,"spread":false}],"spread":false}],"spread":true},{"title":"person","children":[{"title":"001.jpg <span style='color:#111;'> 116.73KB </span>","children":null,"spread":false},{"title":"000.jpg <span style='color:#111;'> 78.88KB </span>","children":null,"spread":false},{"title":"002.jpg <span style='color:#111;'> 19.48KB </span>","children":null,"spread":false},{"title":"004.jpg <span style='color:#111;'> 17.63KB </span>","children":null,"spread":false},{"title":"003.jpg <span style='color:#111;'> 12.80KB </span>","children":null,"spread":false}],"spread":true},{"title":"det","children":[{"title":"zidane.jpg <span style='color:#111;'> 164.99KB </span>","children":null,"spread":false}],"spread":true}],"spread":true},{"title":"main_pose_.py <span style='color:#111;'> 2.74KB </span>","children":null,"spread":false},{"title":"main_keypoints_.py <span style='color:#111;'> 2.67KB </span>","children":null,"spread":false},{"title":"combine","children":[{"title":"gesture_recognition.py <span style='color:#111;'> 3.68KB </span>","children":null,"spread":false},{"title":"human_pose.py <span style='color:#111;'> 9.49KB </span>","children":null,"spread":false},{"title":"base.py <span style='color:#111;'> 3.37KB </span>","children":null,"spread":false},{"title":"__init__.py <span style='color:#111;'> 86B </span>","children":null,"spread":false}],"spread":true},{"title":"tracker","children":[{"title":"kcf.py <span style='color:#111;'> 6.31KB </span>","children":null,"spread":false},{"title":"utils.py <span style='color:#111;'> 3.03KB </span>","children":null,"spread":false}],"spread":true},{"title":"main_classifier_.py <span style='color:#111;'> 1.87KB </span>","children":null,"spread":false}],"spread":false}],"spread":true}]