[{"title":"( 43 个子文件 794.28MB ) 基于opencv的人体姿态估计,实现视频的4分类","children":[{"title":"poose","children":[{"title":".vs","children":[{"title":"ProjectSettings.json <span style='color:#111;'> 37B </span>","children":null,"spread":false},{"title":"slnx.sqlite <span style='color:#111;'> 76.00KB </span>","children":null,"spread":false},{"title":"VSWorkspaceState.json <span style='color:#111;'> 209B </span>","children":null,"spread":false},{"title":"poose","children":[{"title":"v15","children":[{"title":".suo <span style='color:#111;'> 20.50KB </span>","children":null,"spread":false}],"spread":true},{"title":"v16","children":[{"title":".suo <span style='color:#111;'> 51.50KB </span>","children":null,"spread":false}],"spread":true}],"spread":true}],"spread":true},{"title":"poose","children":[{"title":"Tracker.py <span style='color:#111;'> 9.36KB </span>","children":null,"spread":false},{"title":"fall_fight_waalk.mp4 <span style='color:#111;'> 31.86MB </span>","children":null,"spread":false},{"title":"trainy.txt <span style='color:#111;'> 1.59KB </span>","children":null,"spread":false},{"title":"trainx-Copy1.txt <span style='color:#111;'> 3.93MB </span>","children":null,"spread":false},{"title":"photo.jpg <span style='color:#111;'> 156.14KB </span>","children":null,"spread":false},{"title":"dance.mp4 <span style='color:#111;'> 4.41MB </span>","children":null,"spread":false},{"title":"fight1.mp4 <span style='color:#111;'> 135.36MB </span>","children":null,"spread":false},{"title":"fall1.mp4 <span style='color:#111;'> 76.53MB </span>","children":null,"spread":false},{"title":"my_best_model.epoch{epoch_02d}-loss{val_loss_.2f}.hdf5 <span style='color:#111;'> 981.04KB </span>","children":null,"spread":false},{"title":"dance1.mp4 <span style='color:#111;'> 208.53MB </span>","children":null,"spread":false},{"title":"trainy-Copy1.txt <span style='color:#111;'> 1.59KB </span>","children":null,"spread":false},{"title":"stand1.mp4 <span style='color:#111;'> 93.13MB </span>","children":null,"spread":false},{"title":"хоЙшгЕцХЩчиЛ.docx <span style='color:#111;'> 663.17KB </span>","children":null,"spread":false},{"title":"FPS.py <span style='color:#111;'> 1.21KB </span>","children":null,"spread":false},{"title":"poose.ipynb <span style='color:#111;'> 58.31KB </span>","children":null,"spread":false},{"title":"models","children":[{"title":"movenet_multipose_lightning_192x256_FP32.bin <span style='color:#111;'> 18.04MB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_320x320_FP32.bin <span style='color:#111;'> 18.07MB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_256x256_FP32.xml <span style='color:#111;'> 515.12KB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_480x640_FP32.xml <span style='color:#111;'> 515.78KB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_192x192_FP32.bin <span style='color:#111;'> 18.04MB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_256x320_FP32.bin <span style='color:#111;'> 18.06MB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_256x256_FP32.bin <span style='color:#111;'> 18.05MB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_320x320_FP32.xml <span style='color:#111;'> 515.28KB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_480x640_FP32.bin <span style='color:#111;'> 18.17MB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_192x256_FP32.xml <span style='color:#111;'> 515.12KB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_736x1280_FP32.bin <span style='color:#111;'> 18.47MB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_256x320_FP32.xml <span style='color:#111;'> 515.19KB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_192x192_FP32.xml <span style='color:#111;'> 515.09KB </span>","children":null,"spread":false},{"title":"movenet_multipose_lightning_736x1280_FP32.xml <span style='color:#111;'> 515.86KB </span>","children":null,"spread":false}],"spread":false},{"title":"trainx.txt <span style='color:#111;'> 3.93MB </span>","children":null,"spread":false},{"title":"my_best_model-Copy1.epoch{epoch_02d}-loss{val_loss_.2f}.hdf5 <span style='color:#111;'> 981.04KB </span>","children":null,"spread":false},{"title":"__pycache__","children":[{"title":"FPS.cpython-38.pyc <span style='color:#111;'> 1.72KB </span>","children":null,"spread":false},{"title":"Tracker.cpython-38.pyc <span style='color:#111;'> 8.17KB </span>","children":null,"spread":false},{"title":"FPS.cpython-310.pyc <span style='color:#111;'> 1.74KB </span>","children":null,"spread":false}],"spread":false},{"title":"MovenetOpenvino.py <span style='color:#111;'> 15.66KB </span>","children":null,"spread":false},{"title":".ipynb_checkpoints","children":[{"title":"poose-checkpoint.ipynb <span style='color:#111;'> 1.66MB </span>","children":null,"spread":false}],"spread":false},{"title":"output.mp4 <span style='color:#111;'> 64.38MB </span>","children":null,"spread":false},{"title":"walk1.mp4 <span style='color:#111;'> 63.65MB </span>","children":null,"spread":false}],"spread":false}],"spread":true}],"spread":true}]