torchvision-0.21.0+cu118.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
torchvision-0.21.0+cu118.dist-info/LICENSE,,
torchvision-0.21.0+cu118.dist-info/METADATA,,
torchvision-0.21.0+cu118.dist-info/RECORD,,
torchvision-0.21.0+cu118.dist-info/REQUESTED,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
torchvision-0.21.0+cu118.dist-info/WHEEL,,
torchvision-0.21.0+cu118.dist-info/top_level.txt,,
torchvision-0.21.0+cu118.dist-info\LICENSE,sha256=wGNj-dM2J9xRc7E1IkRMyF-7Rzn2PhbUWH1cChZbWx4,1546
torchvision-0.21.0+cu118.dist-info\METADATA,sha256=Ah-IWUWlPnD7pt65EHTM6TY5JCDTFgXUz-IzF4imkA4,6293
torchvision-0.21.0+cu118.dist-info\RECORD,,
torchvision-0.21.0+cu118.dist-info\WHEEL,sha256=KNRoynpGu-d6mheJI-zfvcGl1iN-y8BewbiCDXsF3cY,101
torchvision-0.21.0+cu118.dist-info\top_level.txt,sha256=ucJZoaluBW9BGYT4TuCE6zoZY_JuSP30wbDh-IRpxUU,12
torchvision/_C.pyd,,
torchvision/__init__.py,,
torchvision/__pycache__/__init__.cpython-312.pyc,,
torchvision/__pycache__/_internally_replaced_utils.cpython-312.pyc,,
torchvision/__pycache__/_meta_registrations.cpython-312.pyc,,
torchvision/__pycache__/_utils.cpython-312.pyc,,
torchvision/__pycache__/extension.cpython-312.pyc,,
torchvision/__pycache__/utils.cpython-312.pyc,,
torchvision/__pycache__/version.cpython-312.pyc,,
torchvision/_internally_replaced_utils.py,,
torchvision/_meta_registrations.py,,
torchvision/_utils.py,,
torchvision/cudart64_110.dll,,
torchvision/datasets/__init__.py,,
torchvision/datasets/__pycache__/__init__.cpython-312.pyc,,
torchvision/datasets/__pycache__/_optical_flow.cpython-312.pyc,,
torchvision/datasets/__pycache__/_stereo_matching.cpython-312.pyc,,
torchvision/datasets/__pycache__/caltech.cpython-312.pyc,,
torchvision/datasets/__pycache__/celeba.cpython-312.pyc,,
torchvision/datasets/__pycache__/cifar.cpython-312.pyc,,
torchvision/datasets/__pycache__/cityscapes.cpython-312.pyc,,
torchvision/datasets/__pycache__/clevr.cpython-312.pyc,,
torchvision/datasets/__pycache__/coco.cpython-312.pyc,,
torchvision/datasets/__pycache__/country211.cpython-312.pyc,,
torchvision/datasets/__pycache__/dtd.cpython-312.pyc,,
torchvision/datasets/__pycache__/eurosat.cpython-312.pyc,,
torchvision/datasets/__pycache__/fakedata.cpython-312.pyc,,
torchvision/datasets/__pycache__/fer2013.cpython-312.pyc,,
torchvision/datasets/__pycache__/fgvc_aircraft.cpython-312.pyc,,
torchvision/datasets/__pycache__/flickr.cpython-312.pyc,,
torchvision/datasets/__pycache__/flowers102.cpython-312.pyc,,
torchvision/datasets/__pycache__/folder.cpython-312.pyc,,
torchvision/datasets/__pycache__/food101.cpython-312.pyc,,
torchvision/datasets/__pycache__/gtsrb.cpython-312.pyc,,
torchvision/datasets/__pycache__/hmdb51.cpython-312.pyc,,
torchvision/datasets/__pycache__/imagenet.cpython-312.pyc,,
torchvision/datasets/__pycache__/imagenette.cpython-312.pyc,,
torchvision/datasets/__pycache__/inaturalist.cpython-312.pyc,,
torchvision/datasets/__pycache__/kinetics.cpython-312.pyc,,
torchvision/datasets/__pycache__/kitti.cpython-312.pyc,,
torchvision/datasets/__pycache__/lfw.cpython-312.pyc,,
torchvision/datasets/__pycache__/lsun.cpython-312.pyc,,
torchvision/datasets/__pycache__/mnist.cpython-312.pyc,,
torchvision/datasets/__pycache__/moving_mnist.cpython-312.pyc,,
torchvision/datasets/__pycache__/omniglot.cpython-312.pyc,,
torchvision/datasets/__pycache__/oxford_iiit_pet.cpython-312.pyc,,
torchvision/datasets/__pycache__/pcam.cpython-312.pyc,,
torchvision/datasets/__pycache__/phototour.cpython-312.pyc,,
torchvision/datasets/__pycache__/places365.cpython-312.pyc,,
torchvision/datasets/__pycache__/rendered_sst2.cpython-312.pyc,,
torchvision/datasets/__pycache__/sbd.cpython-312.pyc,,
torchvision/datasets/__pycache__/sbu.cpython-312.pyc,,
torchvision/datasets/__pycache__/semeion.cpython-312.pyc,,
torchvision/datasets/__pycache__/stanford_cars.cpython-312.pyc,,
torchvision/datasets/__pycache__/stl10.cpython-312.pyc,,
torchvision/datasets/__pycache__/sun397.cpython-312.pyc,,
torchvision/datasets/__pycache__/svhn.cpython-312.pyc,,
torchvision/datasets/__pycache__/ucf101.cpython-312.pyc,,
torchvision/datasets/__pycache__/usps.cpython-312.pyc,,
torchvision/datasets/__pycache__/utils.cpython-312.pyc,,
torchvision/datasets/__pycache__/video_utils.cpython-312.pyc,,
torchvision/datasets/__pycache__/vision.cpython-312.pyc,,
torchvision/datasets/__pycache__/voc.cpython-312.pyc,,
torchvision/datasets/__pycache__/widerface.cpython-312.pyc,,
torchvision/datasets/_optical_flow.py,,
torchvision/datasets/_stereo_matching.py,,
torchvision/datasets/caltech.py,,
torchvision/datasets/celeba.py,,
torchvision/datasets/cifar.py,,
torchvision/datasets/cityscapes.py,,
torchvision/datasets/clevr.py,,
torchvision/datasets/coco.py,,
torchvision/datasets/country211.py,,
torchvision/datasets/dtd.py,,
torchvision/datasets/eurosat.py,,
torchvision/datasets/fakedata.py,,
torchvision/datasets/fer2013.py,,
torchvision/datasets/fgvc_aircraft.py,,
torchvision/datasets/flickr.py,,
torchvision/datasets/flowers102.py,,
torchvision/datasets/folder.py,,
torchvision/datasets/food101.py,,
torchvision/datasets/gtsrb.py,,
torchvision/datasets/hmdb51.py,,
torchvision/datasets/imagenet.py,,
torchvision/datasets/imagenette.py,,
torchvision/datasets/inaturalist.py,,
torchvision/datasets/kinetics.py,,
torchvision/datasets/kitti.py,,
torchvision/datasets/lfw.py,,
torchvision/datasets/lsun.py,,
torchvision/datasets/mnist.py,,
torchvision/datasets/moving_mnist.py,,
torchvision/datasets/omniglot.py,,
torchvision/datasets/oxford_iiit_pet.py,,
torchvision/datasets/pcam.py,,
torchvision/datasets/phototour.py,,
torchvision/datasets/places365.py,,
torchvision/datasets/rendered_sst2.py,,
torchvision/datasets/samplers/__init__.py,,
torchvision/datasets/samplers/__pycache__/__init__.cpython-312.pyc,,
torchvision/datasets/samplers/__pycache__/clip_sampler.cpython-312.pyc,,
torchvision/datasets/samplers/clip_sampler.py,,
torchvision/datasets/sbd.py,,
torchvision/datasets/sbu.py,,
torchvision/datasets/semeion.py,,
torchvision/datasets/stanford_cars.py,,
torchvision/datasets/stl10.py,,
torchvision/datasets/sun397.py,,
torchvision/datasets/svhn.py,,
torchvision/datasets/ucf101.py,,
torchvision/datasets/usps.py,,
torchvision/datasets/utils.py,,
torchvision/datasets/video_utils.py,,
torchvision/datasets/vision.py,,
torchvision/datasets/voc.py,,
torchvision/datasets/widerface.py,,
torchvision/extension.py,,
torchvision/image.pyd,,
torchvision/io/__init__.py,,
torchvision/io/__pycache__/__init__.cpython-312.pyc,,
torchvision/io/__pycache__/_load_gpu_decoder.cpython-312.pyc,,
torchvision/io/__pycache__/_video_opt.cpython-312.pyc,,
torchvision/io/__pycache__/image.cpython-312.pyc,,
torchvision/io/__pycache__/video.cpython-312.pyc,,
torchvision/io/__pycache__/video_reader.cpython-312.pyc,,
torchvision/io/_load_gpu_decoder.py,,
torchvision/io/_video_opt.py,,
torchvision/io/image.py,,
torchvision/io/video.py,,
torchvision/io/video_reader.py,,
torchvision/jpeg8.dll,,
torchvision/libjpeg.dll,,
torchvision/libpng16.dll,,
torchvision/libsharpyuv.dll,,
torchvision/libwebp.dll,,
torchvision/models/__init__.py,,
torchvision/models/__pycache__/__init__.cpython-312.pyc,,
torchvision/models/__pycache__/_api.cpython-312.pyc,,
torchvision/models/__pycache__/_meta.cpython-312.pyc,,
torchvision/models/__pycache__/_utils.cpython-312.pyc,,
torchvision/models/__pycache__/alexnet.cpython-312.pyc,,
torchvision/models/__pycache__/convnext.cpython-312.pyc,,
torchvision/models/__pycache__/densenet.cpython-312.pyc,,
torchvision/models/__pycache__/efficientnet.cpython-312.pyc,,
torchvision/models/__pycache__/feature_extraction.cpython-312.pyc,,
torchvision/models/__pycache__/googlenet.cpython-312.pyc,,
torchvision/models/__pycache__/inception.cpython-312.pyc,,
torchvision/models/__pycache__/maxvit.cpython-312.pyc,,
torchvision/models/__pycache__/mnasnet.cpython-312.pyc,,
torchvision/models/__pycache__/mobilenet.cpython-312.pyc,,
torchvision/models/__pycache__/mobilenetv2.cpython-312.pyc,,
torchvision/models/__pycache__/mobilenetv3.cpython-312.pyc,,
torchvision/models/__pycache__/regnet.cpython-312.pyc,,
torchvision/models/__pycache__/resnet.cpython-312.pyc,,
torchvision/models/__pycache__/shufflenetv2.cpython-312.pyc,,
torchvision/models/__pycache__/squeezenet.cpython-312.pyc,,
torchvision/models/__pycache__/swin_transformer.cpython-312.pyc,,
torchvision/models/__pycache__/vgg.cpython-312.pyc,,
torchvision/models/__pycache__/vision_transformer.cpython-312.pyc,,
torchvision/models/_api.py,,
torchvision/models/_meta.py,,
torchvision/models/_utils.py,,
torchvision/models/alexnet.py,,
torchvision/models/convnext.py,,
torchvision/models/densenet.py,,
torchvision/models/detection/__init__.py,,
torchvision/models/detection/__pycache__/__init__.cpython-312.pyc,,
torchvision/models/detection/__pycache__/_utils.cpython-312.pyc,,
torchvision/models/detection/__pycache__/anchor_utils.cpython-312.pyc,,
torchvision/models/detection/__pycache__/backbone_utils.cpython-312.pyc,,
torchvision/models/detection/__pycache__/faster_rcnn.cpython-312.pyc,,
torchvision/models/detection/__pycache__/fcos.cpython-312.pyc,,
torchvision/models/detection/__pycache__/generalized_rcnn.cpython-312.pyc,,
torchvision/models/detection/__pycache__/image_list.cpython-312.pyc,,
torchvision/models/detection/__pycache__/keypoint_rcnn.cpython-312.pyc,,
torchvision/models/detection/__pycache__/mask_rcnn.cpython-312.pyc,,
torchvision/models/detection/__pycache__/retinanet.cpython-312.pyc,,
torchvision/models/detection/__pycache__/roi_heads.cpython-312.pyc,,
torchvision/models/detection/__pycache__/rpn.cpython-312.pyc,,
torchvision/models/detection/__pycache__/ssd.cpython-312.pyc,,
torchvision/models/detection/__pycache__/ssdlite.cpython-312.pyc,,
torchvision/models/detection/__pycache__/transform.cpython-312.pyc,,
torchvision/models/detection/_utils.py,,
torchvision/models/detection/anchor_utils.py,,
torchvision/models/detection/backbone_utils.py,,
torchvision/models/detection/faster_rcnn.py,,
torchvision/models/detection/fcos.py,,
torchvision/models/detection/generalized_rcnn.py,,
torchvision/models/detection/image_list.py,,
torchvision/models/detection/keypoint_rcnn.py,,
torchvision/models/detection/mask_rcnn.py,,
torchvision/models/detection/retinanet.py,,
torchvision/models/detection/roi_heads.py,,
torchvision/models/detection/rpn.py,,
torchvision/models/detection/ssd.py,,
torchvision/models/detection/ssdlite.py,,
torchvision/models/detection/transform.py,,
torchvision/models/efficientnet.py,,
torchvision/models/feature_extraction.py,,
torchvision/models/googlenet.py,,
torchvision/models/inception.py,,
torchvision/models/maxvit.py,,
torchvision/models/mnasnet.py,,
torchvision/models/mobilenet.py,,
torchvision/models/mobilenetv2.py,,
torchvision/models/mobilenetv3.py,,
torchvision/models/optical_flow/__init__.py,,
torchvision/models/optical_flow/__pycache__/__init__.cpython-312.pyc,,
torchvision/models/optical_flow/__pycache__/_utils.cpython-312.pyc,,
torchvision/models/optical_flow/__pycache__/raft.cpython-312.pyc,,
torchvision/models/optical_flow/_utils.py,,
torchvision/models/optical_flow/raft.py,,
torchvision/models/quantization/__init__.py,,
torchvision/models/quantization/__pycache__/__init__.cpython-312.pyc,,
torchvision/models/quantization/__pycache__/googlenet.cpython-312.pyc,,
torchvision/models/quantization/__pycache__/inception.cpython-312.pyc,,
torchvision/models/quantization/__pycache__/mobilenet.cpython-312.pyc,,
torchvision/models/quantization/__pycache__/mobilenetv2.cpython-312.pyc,,
torchvision/models/quantization/__pycache__/mobilenetv3.cpython-312.pyc,,
torchvision/models/quantization/__pycache__/resnet.cpython-312.pyc,,
torchvision/models/quantization/__pycache__/shufflenetv2.cpython-312.pyc,,
torchvision/models/quantization/__pycache__/utils.cpython-312.pyc,,
torchvision/models/quantization/googlenet.py,,
torchvision/models/quantization/inception.py,,
torchvision/models/quantization/mobilenet.py,,
torchvision/models/quantization/mobilenetv2.py,,
torchvision/models/quantization/mobilenetv3.py,,
torchvision/models/quantization/resnet.py,,
torchvision/models/quantization/shufflenetv2.py,,
torchvision/models/quantization/utils.py,,
torchvision/models/regnet.py,,
torchvision/models/resnet.py,,
torchvision/models/segmentation/__init__.py,,
torchvision/models/segmentation/__pycache__/__init__.cpython-312.pyc,,
torchvision/models/segmentation/__pycache__/_utils.cpython-312.pyc,,
torchvision/models/segmentation/__pycache__/deeplabv3.cpython-312.pyc,,
torchvision/models/segmentation/__pycache__/fcn.cpython-312.pyc,,
torchvision/models/segmentation/__pycache__/lraspp.cpython-312.pyc,,
torchvision/models/segmentation/_utils.py,,
torchvision/models/segmentation/deeplabv3.py,,
torchvision/models/segmentation/fcn.py,,
torchvision/models/segmentation/lraspp.py,,
torchvision/models/shufflenetv2.py,,
torchvision/models/squeezenet.py,,
torchvision/models/swin_transformer.py,,
torchvision/models/vgg.py,,
torchvision/models/video/__init__.py,,
torchvision/models/video/__pycache__/__init__.cpython-312.pyc,,
torchvision/models/video/__pycache__/mvit.cpython-312.pyc,,
torchvision/models/video/__pycache__/resnet.cpython-312.pyc,,
torchvision/models/video/__pycache__/s3d.cpython-312.pyc,,
torchvision/models/video/__pycache__/swin_transformer.cpython-312.pyc,,
torchvision/models/video/mvit.py,,
torchvision/models/video/resnet.py,,
torchvision/models/video/s3d.py,,
torchvision/models/video/swin_transformer.py,,
torchvision/models/vision_transformer.py,,
torchvision/nvjpeg64_11.dll,,
torchvision/ops/__init__.py,,
torchvision/ops/__pycache__/__init__.cpython-312.pyc,,
torchvision/ops/__pycache__/_box_convert.cpython-312.pyc,,
torchvision/ops/__pycache__/_register_onnx_ops.cpython-312.pyc,,
torchvision/ops/__pycache__/_utils.cpython-312.pyc,,
torchvision/ops/__pycache__/boxes.cpython-312.pyc,,
torchvision/ops/__pycache__/ciou_loss.cpython-312.pyc,,
torchvision/ops/__pycache__/deform_conv.cpython-312.pyc,,
torchvision/ops/__pycache__/diou_loss.cpython-312.pyc,,
torchvision/ops/__pycache__/drop_block.cpython-312.pyc,,
torchvision/ops/__pycache__/feature_pyramid_network.cpython-312.pyc,,
torchvision/ops/__pycache__/focal_loss.cpython-312.pyc,,
torchvision/ops/__pycache__/giou_loss.cpython-312.pyc,,
torchvision/ops/__pycache__/misc.cpython-312.pyc,,
torchvision/ops/__pycache__/poolers.cpython-312.pyc,,
torchvision/ops/__pycache__/ps_roi_align.cpython-312.pyc,,
torchvision/ops/__pycache__/ps_roi_pool.cpython-312.pyc,,
torchvision/ops/__pycache__/roi_align.cpython-312.pyc,,
torchvision/ops/__pycache__/roi_pool.cpython-312.pyc,,
torchvision/ops/__pycache__/stochastic_depth.cpython-312.pyc,,
torchvision/ops/_box_convert.py,,
torchvision/ops/_register_onnx_ops.py,,
torchvision/ops/_utils.py,,
torchvision/ops/boxes.py,,
torchvision/ops/ciou_loss.py,,
torchvision/ops/deform_conv.py,,
torchvision/ops/diou_loss.py,,
torchvision/ops/drop_block.py,,
torchvision/ops/feature_pyramid_network.py,,
torchvision/ops/focal_loss.py,,
torchvision/ops/giou_loss.py,,
torchvision/ops/misc.py,,
torchvision/ops/poolers.py,,
torchvision/ops/ps_roi_align.py,,
torchvision/ops/ps_roi_pool.py,,
torchvision/ops/roi_align.py,,
torchvision/ops/roi_pool.py,,
torchvision/ops/stochastic_depth.py,,
torchvision/transforms/__init__.py,,
torchvision/transforms/__pycache__/__init__.cpython-312.pyc,,
torchvision/transforms/__pycache__/_functional_pil.cpython-312.pyc,,
torchvision/transforms/__pycache__/_functional_tensor.cpython-312.pyc,,
torchvision/transforms/__pycache__/_functional_video.cpython-312.pyc,,
torchvision/transforms/__pycache__/_presets.cpython-312.pyc,,
torchvision/transforms/__pycache__/_transforms_video.cpython-312.pyc,,
torchvision/transforms/__pycache__/autoaugment.cpython-312.pyc,,
torchvision/transforms/__pycache__/functional.cpython-312.pyc,,
torchvision/transforms/__pycache__/transforms.cpython-312.pyc,,
torchvision/transforms/_functional_pil.py,,
torchvision/transforms/_functional_tensor.py,,
torchvision/transforms/_functional_video.py,,
torchvision/transforms/_presets.py,,
torchvision/transforms/_transforms_video.py,,
torchvision/transforms/autoaugment.py,,
torchvision/transforms/functional.py,,
torchvision/transforms/transforms.py,,
torchvision/transforms/v2/__init__.py,,
torchvision/transforms/v2/__pycache__/__init__.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_augment.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_auto_augment.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_color.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_container.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_deprecated.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_geometry.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_meta.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_misc.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_temporal.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_transform.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_type_conversion.cpython-312.pyc,,
torchvision/transforms/v2/__pycache__/_utils.cpython-312.pyc,,
torchvision/transforms/v2/_augment.py,,
torchvision/transforms/v2/_auto_augment.py,,
torchvision/transforms/v2/_color.py,,
torchvision/transforms/v2/_container.py,,
torchvision/transforms/v2/_deprecated.py,,
torchvision/transforms/v2/_geometry.py,,
torchvision/transforms/v2/_meta.py,,
torchvision/transforms/v2/_misc.py,,
torchvision/transforms/v2/_temporal.py,,
torchvision/transforms/v2/_transform.py,,
torchvision/transforms/v2/_type_conversion.py,,
torchvision/transforms/v2/_utils.py,,
torchvision/transforms/v2/functional/__init__.py,,
torchvision/transforms/v2/functional/__pycache__/__init__.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_augment.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_color.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_deprecated.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_geometry.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_meta.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_misc.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_temporal.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_type_conversion.cpython-312.pyc,,
torchvision/transforms/v2/functional/__pycache__/_utils.cpython-312.pyc,,
torchvision/transforms/v2/functional/_augment.py,,
torchvision/transforms/v2/functional/_color.py,,
torchvision/transforms/v2/functional/_deprecated.py,,
torchvision/transforms/v2/functional/_geometry.py,,
torchvision/transforms/v2/functional/_meta.py,,
torchvision/transforms/v2/functional/_misc.py,,
torchvision/transforms/v2/functional/_temporal.py,,
torchvision/transforms/v2/functional/_type_conversion.py,,
torchvision/transforms/v2/functional/_utils.py,,
torchvision/tv_tensors/__init__.py,,
torchvision/tv_tensors/__pycache__/__init__.cpython-312.pyc,,
torchvision/tv_tensors/__pycache__/_bounding_boxes.cpython-312.pyc,,
torchvision/tv_tensors/__pycache__/_dataset_wrapper.cpython-312.pyc,,
torchvision/tv_tensors/__pycache__/_image.cpython-312.pyc,,
torchvision/tv_tensors/__pycache__/_mask.cpython-312.pyc,,
torchvision/tv_tensors/__pycache__/_torch_function_helpers.cpython-312.pyc,,
torchvision/tv_tensors/__pycache__/_tv_tensor.cpython-312.pyc,,
torchvision/tv_tensors/__pycache__/_video.cpython-312.pyc,,
torchvision/tv_tensors/_bounding_boxes.py,,
torchvision/tv_tensors/_dataset_wrapper.py,,
torchvision/tv_tensors/_image.py,,
torchvision/tv_tensors/_mask.py,,
torchvision/tv_tensors/_torch_function_helpers.py,,
torchvision/tv_tensors/_tv_tensor.py,,
torchvision/tv_tensors/_video.py,,
torchvision/utils.py,,
torchvision/version.py,,
torchvision/zlib.dll,,
torchvision\_C.pyd,sha256=uFwFUNuMOYgUTGkq5e8Urur8p4-HrnMOiFZr1-8vM-Y,5799424
torchvision\__init__.py,sha256=yUybb2eKRBX2LWsDZPAnEpHWtEFcGHS-Xo5r8K-Ric0,3639
torchvision\_internally_replaced_utils.py,sha256=LogrBRyA6oM0joW1_kzgVYicV6ZYCFsszHbcuPxgGSA,1458
torchvision\_meta_registrations.py,sha256=_rNIAEDFkHMmgYGm982-WgSmhA9o2KVJpd8DsoUlQmk,7433
torchvision\_utils.py,sha256=kcSn6P3Vjv5QPgHHNmNu59Mh-NLb5MDlYxvDkNJWrOM,966
torchvision\cudart64_110.dll,sha256=ulwvtSbE7kuyGM6z-l6L_eic5HTzhxH9zOgCVJv5_G8,526848
torchvision\datasets\__init__.py,sha256=Ajw_RJcPNYz-iKvkvQEkgw2P341llDsDgd8cke-bxas,3753
torchvision\datasets\_optical_flow.py,sha256=vpLExp77a-CpVB-jp-WKPL49tGrO5yRQYg9FX1zPiN8,20187
torchvision\datasets\_stereo_matching.py,sha256=z-wTMLgVQZlTfeL9POzk9xbbR-xwcEJqYiYCePqT6Rs,50281
torchvision\datasets\caltech.py,sha256=HwTa-G4rGIzyXXsV9U5cFQoCp-f5BETeP8l078S9d2A,9055
torchvision\datasets\celeba.py,sha256=v8z17O13m_dPWEOp4eAUDHpaPSMYNWEPLx3aQrN13tk,8630
torchvision\datasets\cifar.py,sha256=SrmM2jGfT3jRG8rPj9E6mjT89TgZZu6FLG8JU2DXmWc,5958
torchvision\datasets\cityscapes.py,sha256=UXjwN_IEoNoNLUkWjVAIw7-a87ew1lhK6aVvFx0_IXU,10571
torchvision\datasets\clevr.py,sha256=vliWOSDso7wRX5NtfwWSOK4sL-h0A0UP7tctAie9o3o,3548
torchvision\datasets\coco.py,sha256=ffMMWD4lEElgMPb1V967OTU1RiSt2MPklgh3h3vGOjQ,4289
torchvision\datasets\country211.py,sha256=_iKc9AaBoGm9vWTd9Z2IbdB4hD2KjzKo9Xsa8eDvlfk,2494
torchvision\datasets\dtd.py,sha256=xCxrh3Prnz6Td0D2zp0LY4NupqUN9rihTkb1g8I7mIs,4119
torchvision\datasets\eurosat.py,sha256=9UaXsjqZTo7uOiTy38jKOHX0DXEm4uT8s2BlEYnSaYk,2369
torchvision\datasets\fakedata.py,sha256=BqGViPyIRcvdBNpGhsPl_isrxxOTtlFLyKnLxzBWQP8,2514
torchvision\datasets\fer2013.py,sha256=d1cRDnnrPP8M-DxeguCLnF6vmQMO2QC9LX6Zk8pPfrE,5238
torchvision\datasets\fgvc_aircraft.py,sha256=IlkE04918JZ8of8bwfSPtvfkYafrBIUGUbpg3isfQos,4741
torchvision\datasets\flickr.py,sha256=kRzv6HJiz8TDa7_SR9gc13S1n_4loCz5ike7UhV4S3I,5598
torchvision\datasets\flowers102.py,sha256=pkBqqDqOI4Ef_OB5J8_P8s0dqq_YldY1kkQ0hc3S0No,4755
torchvision\datasets\folder.py,sha256=pQ2bOqR0pURHGT_QSvHRhcUIZIYrebB49Q4v8uWv5fM,13256
torchvision\datasets\food101.py,sha256=jhWKGMqqkIMDHq_dQocQOgYWHdZ6pYKsy9cDB7MPiI0,3845
torchvision\datasets\gtsrb.py,sha256=TVMesJ3jzVk1tg4ecJyiun5lQ0PKh7lMZWLkZbkVbUY,3888
torchvision\datasets\hmdb51.py,sha256=nAfL2TmZMNAuxv6hN1X3jNP-DmBEF5ls5YnjHo64TIs,6123
torchvision\datasets\imagenet.py,sha256=ftqZ3qfRBuL-fAAXFjNwKzN92KdT_vS-PUqIlnFUMfA,8910
torchvision\datasets\imagenette.py,sha256=E8zhlq17Z3Mln-AU8Wo-vYJyhmaRxYHatz5sSdG6KJE,4368
torchvision\datasets\inaturalist.py,sha256=8ZivtVYu8nOKiaW1tgNjq0hW2pcub03vMtZ19sa6XIA,10115
torchvision\datasets\kinetics.py,sha256=7aRbG-PXkhm9jahFHq6aIyw-QKdyUJqVdvV7UjmT1_k,10115
torchvision\datasets\kitti.py,sha256=q0otSanUygTFq4P2T3xKft1Gv-BPQXw-xNNMfOmDOzc,5795
torchvision\datasets\lfw.py,sha256=Ok_XyQJSrYYWWO-Yu1oUfstv-DyjmbSXEKTIEbV0__A,10756
torchvision\datasets\lsun.py,sha256=TomHtE2AGd3TISKGcMuGvJXSeVtH6x6nNq-5B-1kl0M,5911
torchvision\datasets\mnist.py,sha256=EkG4m89O_lY-LglFtIUTYyF9HiNcyeMmgkZ0VTBxkak,22311
torchvision\datasets\moving_mnist.py,sha256=wAzvEmxI0skZFYxUu9gpb7iuV09uc0I5aOjL2JzjgN8,3738
torchvision\datasets\omniglot.py,sha256=EKIFQBCIVzqaDmDdmW09Vh_arqNvU5WKM7WksDMdpE0,4194
torchvision\datasets\oxford_iiit_pet.py,sha256=3RbU_UxvVqifmqzKVMWFf-dttf53GpbndN7518obVsI,5651
torchvision\datasets\pcam.py,sha256=8tWbJcxo7lKYTToISFpL6_OSbxeYEqU2QsaY6OYjRpM,5419
torchvision\datasets\phototour.py,sha256=DI7oqvaijkoWMpogfvkK-L5QAY8-H1bvN9IBVYR25zA,8098
torchvision\datasets\places365.py,sha256=xWT7FCjjW7LwzLZudQJ_9rE9M5x9mdkIoTibId8lTrg,7238
torchvision\datasets\rendered_sst2.py,sha256=_guhu305oAtxCH8KcV4rUwP5CUBucFCYfx--lpBjRK8,3683
torchvision\datasets\samplers\__init__.py,sha256=yX8XD3h-VwTCoqF9IdQmcuGblZxvXb2EIqc5lPqXXtY,164
torchvision\datasets\samplers\clip_sampler.py,sha256=ERp9c0O3ZGQNR3lLXvZXWjfJkXayYym8bhYfOm_MNKI,6416
torchvision\datasets\sbd.py,sha256=aSgEpJSgnh9uUp6xEGXTGRIERSY6hdT5RWpw3dnCap0,5540
torchvision\datasets\sbu.py,sha256=6eKb_eNqVM2dgYfNB4Y2VjvrmZ2YMuLS0rfQ0YDuwhY,4193
torchvision\datasets\semeion.py,sha256=6TvuVqBRvGPXj7SGPm3Q2QnnFieH8d7gOe-OMskFydg,3180
torchvision\datasets\stanford_cars.py,sha256=N3o61NK8hY-Gbxw-vO3XzRW6DfyzvmVHVwafhWRHSBA,5012
torchvision\datasets\stl10.py,sha256=dGiWMjM2cLaAcj5iQ7oJUVv43IL8SUNKX0Up6HRj-8c,7408
torchvision\datasets\sun397.py,sha256=yIbrKH6CfuJ3l8yjQGsZCakQGPcbIxNGb6h64bQCDU4,2859
torchvision\datasets\svhn.py,sha256=nP-62KxLkLD6n84lw5lgyblHyRFWC5NYiBLoGgeID5Y,4958
torchvision\datasets\ucf101.py,sha256=zM9EY6Clqdvt45PZyL0dZvnlLn-5oyCD6v924bENlPE,5664
torchvision\datasets\usps.py,sha256=vaztMPU18onO_7tWXUWNQ0QsPPYoxGB-3a7YiA3pYrs,3596
torchvision\datasets\utils.py,sha256=CMvpluClWFS6KIPczX18ruPwGC-5q7uNAC75H8ak6xk,16465
torchvision\datasets\video_utils.py,sha256=gUazc9gv4wgFuJ1N47e2c4Wi86HWhR9O09HcwVraeLA,17632
torchvision\datasets\vision.py,sha256=5N69xiFsNdiqjVzP-QP3-PtswDUNsKSeyy_FzXTDt_4,4360
torchvision\datasets\voc.py,sha256=7GhzVyU3iWbBzFyb1zdj-9xxCSLmQCdihPO_o85SdqA,9059
torchvision\datasets\widerface.py,sha256=MU382b0UIeX3BaBtwin5BKWoLnTXOPhIR89vrNmLrsU,8460
torchvision\extension.py,sha256=0A4efQ6V8RlQcMMtDpK74VIBHpDv4icjkkOc-EokPHw,3233
torchvision\image.pyd,sha256=aFk1WyE9-2O5VcRfYvaqV-AvyxHPGDsSW0_ob57I1ZM,203264
torchvision\io\__init__.py,sha256=gzG3Qr6eYLnwemcJFTmWwL8g4Bfd3K9BRVydqo2EnY4,1656
torchvision\io\_load_gpu_decoder.py,sha256=vUWhEXG1Dq3pIJHd8avo4o333Z7zx2DN7FOzIxuGluE,186
torchvision\io\_video_opt.py,sha256=vqC3mJSfGHaBnc9k3OhukpCBQsupvvUVnwtT_BxSCKw,20969
torchvision\io\image.py,sha256=2J0_nN-ZICmw2p8jMPMgQtYM1AOxcksyrD_zPUsyqnA,22229
torchvision\io\video.py,sha256=T-5q9fkppCbO4S4cTzkncA_RrFn6YBUycJOJe-HMs48,18491
torchvision\io\video_reader.py,sha256=m9rHePIP1mL47nwxWK6W1ZFfcY-gAZIOr8HLsSzX7yY,11983
torchvision\jpeg8.dll,sha256=aM-Kj2MkrdHI0gkgpHfh86_icuM26XiMu6gyMGeuKig,552448
torchvision\libjpeg.dll,sha256=85Tk07ezKBsmEjt5H6Yu53aO-ckgPdT6CxG6C-VYyx8,284944
torchvision\libpng16.dll,sha256=nPxu7uIrOxgpEXCLUyjP7rQBCghBVPsL-h8OxeArvc0,192512
torchvision\libsharpyuv.dll,sha256=6Lz4Enzhcx9UiWUBCycwqTHTi_D-35YOM-p8zbwB2QM,34576
torchvision\libwebp.dll,sha256=hN1hG6LP-7hh6DCpbYtX82QTWG0KT2EFtSaX2wrJqOc,389392
torchvision\models\__init__.py,sha256=6QlTJfvjKcUmMJvwSapWUNFXbf2Vo15dVRcBuNSaYko,888
torchvision\models\_api.py,sha256=RJurpplK_q7teeUlnqlMuTTxAorqIEhEcuQydyhzF3s,10331
torchvision\models\_meta.py,sha256=2NSIICoq4MDzPZc00DlGJTgHOCwTBSObSTeRTh3E0tQ,30429
torchvision\models\_utils.py,sha256=X7zduE90fkek8DjukzyENOcZ0iop03R0LIxC_FuAazk,11149
torchvision\models\alexnet.py,sha256=XcldP2UuOkdOUfdxitGS8qHzLH78Ny7VCzTzKsaWITU,4607
torchvision\models\convnext.py,sha256=mML3XALGvQGe3aYkmz6dbC7Wus7ntZWr7IVihWjLV8M,15740
torchvision\models\densenet.py,sha256=Wxvecj8b09Uy2FjKun7ZbagpDmHll8L2c3GbM802Ivs,17273
torchvision\models\detection\__init__.py,sha256=D4cs338Z4BQn5TgX2IKuJC9TD2rtw2svUDZlALR-lwI,175
torchvision\models\detection\_utils.py,sha256=m9bowqjuYiR9A7HI7wJAK_kgFrUYlKSukXOEwuUtRpA,22667
torchvision\models\detection\anchor_utils.py,sha256=TQKWOKDFALsTmYw_BMGIDlT9mNQhukmgnNdFuRjN49w,12127
torchvision\models\detection\backbone_utils.py,sha256=BBKVxCyAY9Q8JYOLQ3oAbWJxcjn5HAbPeAcDb-5JoVA,10792
torchvision\models\detection\faster_rcnn.py,sha256=rhMj68KQ4uDeFKT30PVy2xlqLG43u6tA7i9e4MEM15U,37825
torchvision\models\detection\fcos.py,sha256=3qi44CnUveI7W3Uu7u39fg7H5-2JT1C97dEgdSORLog,35010
torchvision\models\detection\generalized_rcnn.py,sha256=nLVj2o8yr6BW1MN12u30QuFbvtsdllEbUlbNH-dO-G0,4861
torchvision\models\detection\image_list.py,sha256=IzFjxIaMdyFas1IHPBgAuBK3iYJOert5HzGurYJitNk,808
torchvision\models\detection\keypoint_rcnn.py,sha256=gzdI2mw2ububcgBhO8U6u17U_hgAEIADtr45aT5iPt4,22421
torchvision\models\detection\mask_rcnn.py,sha256=HZyuWNdOw8WIdO05-8obUED1A0CVuFd5E21KCzzCfFE,27303
torchvision\models\detection\retinanet.py,sha256=aWY__9rz8r0bT2Lvl0P5llW2hE42s6vNOljaTF4hqOU,38203
torchvision\models\detection\roi_heads.py,sha256=f_Lde69JHugGAaiGWh6ugJ9WUTT8f7InxPry_MZxgZY,34698
torchvision\models\detection\rpn.py,sha256=z4ezvg1XS9uNq_3jIYXzIRnsAhuGClZ5AJ5P0NDziN8,16226
torchvision\models\detection\ssd.py,sha256=QyWW7IihpNG4_SWeVo1-X5J9kDJrasd794-AHIcDxQY,29661
torchvision\models\detection\ssdlite.py,sha256=DRkN07LP7WTKt6Q5WIOraQR37Unx7XIiB70UdhmDCcE,13550
torchvision\models\detection\transform.py,sha256=gou0mGGPzLncV2ZMHkTTkw6UO0fUoJii6Dr2N0PNA_Y,12508
torchvision\models\efficientnet.py,sha256=tW6BpsBProhN6b1o1_sHSikSoqDupQRgdgRETiYjcAY,44221
torchvision\models\feature_extraction.py,sha256=mHanRc9yimz90mj9y2rEBKmd0lL5ayfsAP8VpgDln00,28540
torchvision\models\googlenet.py,sha256=oxCJp5dklEoEWu6VwHSgX-pifaOhZ4gH9bz71P-hE-o,13151
torchvision\models\inception.py,sha256=l9tutwO7KNVa5nfdl1_5f-6lJzQ-NSOCzXPArDILeAA,19329
torchvision\models\maxvit.py,sha256=vHRyAGJ8FsPX_bYZWmmm-6mG6NUq3cICHj00rLd8vow,32929
torchvision\models\mnasnet.py,sha256=PTSF4DchbFgtOchd9kgoPKCcfKH6NC4WfHa5bv4jZ58,18008
torchvision\models\mobilenet.py,sha256=alrEJwktmXVcCphU8h2EAJZX0YdKfcz4tJEOdG2BXB8,217
torchvision\models\mobilenetv2.py,sha256=eVil23yP4f-DBUhjKt73mao84zEiF5Y01wHNwh8HCdM,9970
torchvision\models\mobilenetv3.py,sha256=g1Ul1RkohPHhXfsecg2r0W7NVlEgVUc7X_uT1bxxrTQ,16702
torchvision\models\optical_flow\__init__.py,sha256=uuRFAdvcDobdAbY2VmxEZ7_CLH_f5-JRkCSuJRkj4RY,21
torchvision\models\optical_flow\_utils.py,sha256=PRcuU-IB6EL3hAOLiyC5q-NBzlvIKfhSF_BMplHbzfY,2125
torchvision\models\optical_flow\raft.py,sha256=o9wJ3jZH9EWwJl7fQYeWSdeXPMKYOZ0Zwm-hhcequVk,40942
torchvision\models\quantization\__init__.py,sha256=YOJmYqWQTfP5P2ypteZNKQOMW4VEB2WHJlYoSlSaL1Y,130
torchvision\models\quantization\googlenet.py,sha256=P92cacoKVTV4cDoaNkRevLx1daGH5DkPfUwPDMuOXO0,8290
torchvision\models\quantization\inception.py,sha256=TXz2hRpSvh6zYP398MsXTYQMnqYgnYnq5wyG6xr5Nhk,11088
torchvision\models\quantization\mobilenet.py,sha256=alrEJwktmXVcCphU8h2EAJZX0YdKfcz4tJEOdG2BXB8,217
torchvision\models\quantization\mobilenetv2.py,sha256=g2z3HPQ0MXFCuMV5TpLPRdO99m3wC_3d0ukUanXaJHo,6037
torchvision\models\quantization\mobilenetv3.py,sha256=l9g1AwKiU35XKuobXo-UPe5MqRKC1kgN7xgWWim4qr4,9467
torchvision\models\quantization\resnet.py,sha256=azvn1vwebP22qryYGNgFLMviGjHklXOX7xd4C2cggUo,18423
torchvision\models\quantization\shufflenetv2.py,sha256=7k9MLRLzjP3vke-e0Ai_cnA-j15KGQq5yDcs_ELXUg8,17311
torchvision\models\quantization\utils.py,sha256=Ij88l6toyO8MQi1w512Jt-yQ2Q9hK75-Z2SOjIzS6Zw,2109
torchvision\models\regnet.py,sha256=NbsA3RO7ka7k9fwYyVZ5wvvtJUuhXqyX76aGIoIujGE,65124
torchvision\models\resnet.py,sha256=AXWWl7XlkSQpUCsv8lCaWeDuYaq-KyOLXmBe0R8rv58,39917
torchvision\models\segmentation\__init__.py,sha256=TLL2SSmqE08HLiv_yyIWyIyrvf2xaOsZi0muDv_Y5Vc,69
torchvision\models\segmentation\_utils.py,sha256=yFeyBa5_Pyv1UQ_2N64XMRgTYsxifwzDd-VRP-vmIGM,1234
torchvision\models\segmentation\deeplabv3.py,sha256=MkmYEm1dF4afQEYXmFfxLAcqioiT5uWKYGSXCccIYh4,15405
torchvision\models\segmentation\fcn.py,sha256=mQ1Wi4S9j5G6OQbNciuxNwVbJ6e9miTzIWj6mUF5JwA,9205
torchvision\models\segmentation\lraspp.py,sha256=yx_b3PJsH5e0F3TqiGDhEnbXCGTdNX2iIxsKvNenM0s,7821
torchvision\models\shufflenetv2.py,sha256=VEGsTNNTdqdu8m7I62zQuJK_5CkET-0Y4ixYBJ-QBCs,15852
torchvision\models\squeezenet.py,sha256=Dha-ci350KU15D0LS9N07kw6MlNuusUHSBnC83Ery_E,8986
torchvision\models\swin_transformer.py,sha256=-Q9Kd1qzsD7vL3u137Q4MoHSTwzA6QFcweaF0zCWmUk,40370
torchvision\models\vgg.py,sha256=Qt9r5sFoY-oPdwP4De61jqSVe9XUZLXK47r9yVDQ33M,19736
torchvision\models\video\__init__.py,sha256=xHHR5c6kP0cMDXck7XnXq19iJL_Uemcxg3OC00cqE6A,97
torchvision\models\video\mvit.py,sha256=xIK4nCOuJWXQjoX8NzcouwzyTkIkcFug3yiu0a5-Dk8,33895
torchvision\models\video\resnet.py,sha256=JOP7FDfUOfQQP-jEYUvzSIOr9Iaexl2L3iUh-rzrp90,17274
torchvision\models\video\s3d.py,sha256=Rn-iypP13jrETAap1Qd4NY6kkpYDuSXjGkEKZDOxemI,8034
torchvision\models\video\swin_transformer.py,sha256=M74P2v4lVKM6zgwoeFn_njppZB2l-gAjuGVvzzESKpU,28431
torchvision\models\vision_transformer.py,sha256=GE-_-dlFJQPTnONe4qrzYOYp-wavPOrFPCo9krM39Vg,33000
torchvision\nvjpeg64_11.dll,sha256=gVFLDPTckuScHJN99rGS99ttPYU5hldp1Q2TJq2-Nz8,3862528
torchvision\ops\__init__.py,sha256=7wibGxcF1JHDviSNs9O9Pwlc8dhMSFfZo0wzVjTFnAY,2001
torchvision\ops\_box_convert.py,sha256=glF6sulLzaw_KG36wg0CHWt0ef62BnkjokbqQnBUMsU,2490
torchvision\ops\_register_onnx_ops.py,sha256=g4M5Fp7n_5ZTzIQcUXvEct3YFlUMPNVSQQfBP-J0eQQ,4288
torchvision\ops\_utils.py,sha256=xFrLnLhKDHiG2TN39tUWY-MJbLEPka6dkaVVJFAN7-8,3736
torchvision\ops\boxes.py,sha256=C76mPf_218F5BHrpuxj1b31OsKKnPzDPnGBiiHe0Wto,17044
torchvision\ops\ciou_loss.py,sha256=Qzm89C82ehX-YvYBPLPRPhbJZdr3itizxuxrT7MLi9o,2834
torchvision\ops\deform_conv.py,sha256=DuIosFDK3tsY5RlHU7mez5x1p08IQai9WG14z3S0gzU,7185
torchvision\ops\diou_loss.py,sha256=6IebWlMYc_2YnbG36niDXgM16vxajSKRfiusEuUJwpQ,3456
torchvision\ops\drop_block.py,sha256=ZkIzM1b3v5_U7z0eavzaNpN7IBq0N4ZNwwvWArispwg,6010
torchvision\ops\feature_pyramid_network.py,sha256=Jts5mzUJX3EarcAQU5MDUe0a5Sgn5YjUstaW2JQpgEE,8952
torchvision\ops\focal_loss.py,sha256=lS5FqgLFuDKlpm0sk5V1VgIA6LFAdJUXQaPi35nEDoU,2319
torchvision\ops\giou_loss.py,sha256=xB_RlES28k_A6iH2VqWAPBQkiT_zkEwdtREDGR_nVJM,2772
torchvision\ops\misc.py,sha256=niQnKPuifQzVXAWnnf6TkVsgetqyetjZ6lq3wi2tsZw,13892
torchvision\ops\poolers.py,sha256=sfgcZWh2dIo9UY437CnpAHdxqPQhuvjNPYzhIKlAIPE,12247
torchvision\ops\ps_roi_align.py,sha256=6_kmnE6z_3AZZ1N2hrS_uK3cbuzqZhjdM2rC50mfYUo,3715
torchvision\ops\ps_roi_pool.py,sha256=2JrjJwzVtEeEg0BebkCnGfq4xEDwMCD-Xh915mvNcyI,2940
torchvision\ops\roi_align.py,sha256=jOBOfcpY5RcJkJ6J8VF-OMSvBaYWMxZLKrv1Oc6Dmag,11585
torchvision\ops\roi_pool.py,sha256=cN7rSCQfpUzETvP8SjPDl1yfXjbxg9I-tXnHbvAKya8,3015
torchvision\ops\stochastic_depth.py,sha256=9T4Zu_BaemKZafSmRwrPCVr5aaGH8tmzlsQAZO-1_-Y,2302
torchvision\transforms\__init__.py,sha256=WCNXTJUbJ1h7YaN9UfrBSvt--ST2PAV4sLICbTS-L5A,55
torchvision\transforms\_functional_pil.py,sha256=a9b6BkJnMAL2tGLknE-iQY8TQcBcaPSK54Kqml8WDsM,12463
torchvision\transforms\_functional_tensor.py,sha256=0-cJE2Ns0WpLc0hZzdC_8R3vwA8dAxdRunEnjbF6YTY,34901
torchvision\transforms\_functional_video.py,sha256=c4BbUi3Y2LvskozFdy619piLBd5acsjxgogYAXmY5P8,3971
torchvision\transforms\_presets.py,sha256=vVYoqYy3dY3j9MuLHeS8B_q5qrBJFmI1ggTBsobUeYs,8726
torchvision\transforms\_transforms_video.py,sha256=ub2gCT5ELiK918Bq-Pp6mzhWrAZxlj7blfpkA8Dhb1o,5124
torchvision\transforms\autoaugment.py,sha256=UD8UBlT4dWCIQaNDUDQBtc0osMHHPQluLr7seZJr4cY,28858
torchvision\transforms\functional.py,sha256=e_eEdiTITSrhU2aNFRp8iqenfyFxevsgrmFBAZKM_TU,69441
torchvision\transforms\transforms.py,sha256=dcHYDZIbYOW6sOywH41uOiBd9tPfEgOdfFbSzqONtfA,87700
torchvision\transforms\v2\__init__.py,sha256=d5VraJDV5PXPv5mifqUTdQx2z1p6M3ks3RNuRc-ztfk,1605
torchvision\transforms\v2\_augment.py,sha256=iNI_1WvGkvcRqoGHRb-8VMuzUXUUFEQefCq6gqLQha4,16523
torchvision\transforms\v2\_auto_augment.py,sha256=rwzHS7CZvt2r21EuIzSKW4GCrsNddYooTRXFkH3bOOo,32652
torchvision\transforms\v2\_color.py,sha256=jM6PtB76IQrhpr5QsF-GerSmTVr_FiW52xahXfsY4Fg,17354
torchvision\transforms\v2\_container.py,sha256=E-8TvTF_qBqC6aLnlK5mnp3V27oQOxhy1X7PZfQAD5w,6229
torchvision\transforms\v2\_deprecated.py,sha256=ACthGTn1ciZtp9py8njxypGEJgObcjfYJ4H4MFY8OPw,1996
torchvision\transforms\v2\_geometry.py,sha256=hTbRiUyEFdRtMB0tcuKJugnBXO7fIAl6ivd4CV0ORI0,69001
torchvision\transforms\v2\_meta.py,sha256=6CK1TP1vADvSGqrMSJcuXTRLxPMHNFmWn3Pdrlg74oc,1439
torchvision\transforms\v2\_misc.py,sha256=cUntkv21XAZhMRZY_X3Csm2Ja6fTEQs7_FHTiS4WXSQ,19553
torchvision\transforms\v2\_temporal.py,sha256=IgEq3Crm49frD7RHmv5xa9ZyQUIP9kXwVW28icQRdgY,931
torchvision\transforms\v2\_transform.py,sha256=O8cudZScPVaemfAEvW9IYnDrXCst9DdoIDd72azfqKc,9561
torchvision\transforms\v2\_type_conversion.py,sha256=2mcO70MRfpDmHMWdjI2nCdEpqxwbT0zrxy0VTuijfuE,2940
torchvision\transforms\v2\_utils.py,sha256=ftioCRuUHm9eBnODdx5dBP7ZonPiTIrzG9_dwJ-RUSY,9072
torchvision\transforms\v2\functional\__init__.py,sha256=LGu6ZNvdGaD7gED4Mi7x-3ES1D9FnN3xraVEb_p94eI,3699
torchvision\transforms\v2\functional\_augment.py,sha256=oFPymp04zhQpTQR3O-5XUJHDnQW6-pfwWJKoP-dpWkM,3579
torchvision\transforms\v2\functional\_color.py,sha256=Emetqsc4q5hlwTtBq-wlsnOZScsPqa8VaA1MndrILto,30983
torchvision\transforms\v2\functional\_deprecated.py,sha256=-X7agTXp-JnbFpsp3xoVk1eZr7OyT-evBt0nlULAR40,825
torchvision\transforms\v2\functional\_geometry.py,sha256=SnFMvBIgzuRBkYzGY5e-rDvcn8CXyHvKQhXzJBzq5Ag,89897
torchvision\transforms\v2\functional\_meta.py,sha256=b_MF4SQrmZNVpcN8X-8edP-CEmQ241MKDWpQTG4pmbI,10826
torchvision\transforms\v2\functional\_misc.py,sha256=q30X0WtS3uYrPoTktS-2458NFdEP1j-PG6r69HwEAtA,17937
torchvision\transforms\v2\functional\_temporal.py,sha256=tSRkkqOqUQ0QXjENF82F16El1-J0IDoFKIH-ss_cpC4,1163
torchvision\transforms\v2\functional\_type_conversion.py,sha256=oYf4LMgiClvEZwwc3WbKI7fJ-rRFhDrVSBKiPA5vxio,896
torchvision\transforms\v2\functional\_utils.py,sha256=3T5iFgq8whHQbk7duizYmoxspH6mtkP-L7ku627zfBY,5620
torchvision\tv_tensors\__init__.py,sha256=7UBIZbraVyhIO-ZCeKtD59if85XKkAQ4njjS6RjEfDg,1544
torchvision\tv_tensors\_bounding_boxes.py,sha256=t6HTK1MJ1xiiJv0miMjGezQotEw8gQNUObPFMCh0HYU,4596
torchvision\tv_tensors\_dataset_wrapper.py,sha256=pgcasnVwGiQb0mmEqPkIgycI-4AV703xbBRwUFbMo9k,25171
torchvision\tv_tensors\_image.py,sha256=wI0FrnbHUOzaIyna_Lyv5n2TGmSWwjh-HfVTC4LZklo,1987
torchvision\tv_tensors\_mask.py,sha256=soZV5aMAtzlW32Wb52ApagodZ4butgQS8C-eCIirLu0,1517
torchvision\tv_tensors\_torch_function_helpers.py,sha256=U7r-QG2jKV_KYeUQJ2hKPlYRLwqz_xhEs7_L1oXpBvM,2348
torchvision\tv_tensors\_tv_tensor.py,sha256=v-dVm-ZZs4fdT6TVcAeX7KnAJpudVUf55VqdtQtgthE,6380
torchvision\tv_tensors\_video.py,sha256=KtcHdpsm81uhx451oNs6bqFeuUw6G9XjqltDeUJNvzc,1453
torchvision\utils.py,sha256=L0XWXodEwJEy0OxpUlz5W6SYW_jLG2NwENqnwb54pQc,27972
torchvision\version.py,sha256=9wXzZAenmuXIjwwE8m56_wXGC5AblAHiMMhzqEZNQfg,208
torchvision\zlib.dll,sha256=jb_W73N0qDEVi93Mt549VmXpYlyBr1V_FbQVC3h39oc,99608
