admin管理员组文章数量:1345064
I'm working on implementing image segmentation using my own custom TFLite model, following the code example from MediaPipe. Here's my code:
options = vision.ImageSegmenterOptions(
base_options=base_options,
running_mode=mp.tasks.vision.RunningMode.IMAGE,
output_confidence_masks=True,
output_category_mask=False
)
mp_image = mp.Image.create_from_file(image_path)
with vision.ImageSegmenter.create_from_options(options) as segmenter:
segmentation_result = segmenter.segment(mp_image)
output_mask = segmentation_result.confidence_masks[0]
I've encountered two issues with the above code:
The model has two outputs:
Output 0: Name = Identity0, Shape = [1, 1], Type = numpy.float32
Output 1: Name = Identity1, Shape = [1, x, y, z], Type = numpy.float32 (where x * y * z == image_width * image_height * image_channel=1)
How can I retrieve both outputs instead of just one?
The confidence_masks values are almost identical (min/max = 0.0701157/0.070115715), which seems unusual. The original image contains a person, and the output is correct when using my custom TFLite model with tf.lite.Interpreter.get_tensor().
I know that many frameworks support models with multiple inputs and outputs, so I'm confused about what I might be missing. Here are my specific questions:
- Do I need to add special metadata to the TFLite model file?
- How should I modify the original MediaPipe code to handle multiple outputs?
I'm working on implementing image segmentation using my own custom TFLite model, following the code example from MediaPipe. Here's my code:
options = vision.ImageSegmenterOptions(
base_options=base_options,
running_mode=mp.tasks.vision.RunningMode.IMAGE,
output_confidence_masks=True,
output_category_mask=False
)
mp_image = mp.Image.create_from_file(image_path)
with vision.ImageSegmenter.create_from_options(options) as segmenter:
segmentation_result = segmenter.segment(mp_image)
output_mask = segmentation_result.confidence_masks[0]
I've encountered two issues with the above code:
The model has two outputs:
Output 0: Name = Identity0, Shape = [1, 1], Type = numpy.float32
Output 1: Name = Identity1, Shape = [1, x, y, z], Type = numpy.float32 (where x * y * z == image_width * image_height * image_channel=1)
How can I retrieve both outputs instead of just one?
The confidence_masks values are almost identical (min/max = 0.0701157/0.070115715), which seems unusual. The original image contains a person, and the output is correct when using my custom TFLite model with tf.lite.Interpreter.get_tensor().
I know that many frameworks support models with multiple inputs and outputs, so I'm confused about what I might be missing. Here are my specific questions:
- Do I need to add special metadata to the TFLite model file?
- How should I modify the original MediaPipe code to handle multiple outputs?
1 Answer
Reset to default 0Why do you have output_category_mask=False
and are expecting 2 outputs ? You are specifically asking the model to only return 1 output.
Please check the documentation and source code.
output_confidence_masks:
Whether to output confidence.
output_category_mask:
Whether to output category mask.
本文标签: pythonmediapipe Image Segmentation with DualOutput TFLite ModelStack Overflow
版权声明:本文标题:python - mediapipe Image Segmentation with Dual-Output TFLite Model - Stack Overflow 内容由网友自发贡献,该文观点仅代表作者本人, 转载请联系作者并注明出处:http://www.betaflare.com/web/1743788402a2539109.html, 本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌抄袭侵权/违法违规的内容,一经查实,本站将立刻删除。
发表评论