Assigned
Status Update
Comments
po...@gmail.com <po...@gmail.com> #2
I have informed our engineering team of this feature request. There is currently no ETA for its implementation.
A current workaround would be to check the returned "boundingPoly" [1] "vertices" for the returned "textAnnotations". If the calculated rectangle's heights > widths, than your image is sideways.
[1]https://cloud.google.com/vision/reference/rest/v1/images/annotate#boundingpoly
A current workaround would be to check the returned "boundingPoly" [1] "vertices" for the returned "textAnnotations". If the calculated rectangle's heights > widths, than your image is sideways.
[1]
go...@gmail.com <go...@gmail.com> #3
I also need this problem solved :)
nn...@google.com <nn...@google.com> #4
same :D
sh...@gmail.com <sh...@gmail.com> #5
+1
ri...@gmail.com <ri...@gmail.com> #6
+1
ll...@gmail.com <ll...@gmail.com> #7
This needs more attention. It's not just a display issue as described in the report. The co-ordinates returned in 'boundingPoly' are incorrect if the image was taken on a phone. All the x points should be y and vice versa.
The workaround does not make sense as "boundingPoly" [1] "vertices" for "textAnnotations" does not indicate the image dimensions - it indicates the dimensions of the relevant text block inside the image.
The workaround does not make sense as "boundingPoly" [1] "vertices" for "textAnnotations" does not indicate the image dimensions - it indicates the dimensions of the relevant text block inside the image.
ah...@gmail.com <ah...@gmail.com> #8
+1
ar...@asate.com <ar...@asate.com> #9
Would be great if this could be implemented.
ao...@gmail.com <ao...@gmail.com> #10
+1
vh...@gmail.com <vh...@gmail.com> #11
+1
ri...@gmail.com <ri...@gmail.com> #12
+1
ji...@gmail.com <ji...@gmail.com> #13
+1.
di...@gmail.com <di...@gmail.com> #14
The rotation information should already be available, basically the order of the bounding box vertices encode that rotation information:
https://cloud.google.com/vision/docs/reference/rest/v1/images/annotate#block
Could you please test and see if that works for your case?
Could you please test and see if that works for your case?
ri...@gmail.com <ri...@gmail.com> #15
Each bounding box on the page can have a different orientation, so it can be frustrating to figure out.
za...@gmail.com <za...@gmail.com> #16
+1
ri...@gmail.com <ri...@gmail.com> #17
"A current workaround would be to check the returned "boundingPoly" [1] "vertices" for the returned "textAnnotations". If the calculated rectangle's heights > widths, than your image is sideways. " - The proposed workaround does not work if the image needs to be rotated at 180 degrees. Any ETA after 2y and 1 day?
mo...@umail.uom.ac.mu <mo...@umail.uom.ac.mu> #18
+1
ri...@gmail.com <ri...@gmail.com> #19
+1
ka...@gmail.com <ka...@gmail.com> #20
+1
it...@gmail.com <it...@gmail.com> #21
+1
n8...@gmail.com <n8...@gmail.com> #22
Any ETA after such long time.
to...@gmail.com <to...@gmail.com> #23
We just prioritized this work and tentatively aiming for August for releasing various OCR improvements including orientation detection.
m....@program-ace.com.ua <m....@program-ace.com.ua> #24
Great stuff, really good to hear, this will have some really useful applications!
[Deleted User] <[Deleted User]> #25
Just to gather some details, this feature request is for *page* level orientation information? Like a enum UP DOWN LEFT RIGHT. Is orientation also desired for individual blocks, paragraphs, words, symbols, etc?
Also, this new orientation info will only apply to the DOCUMENT_TEXT_DETECTION feature.
Also, this new orientation info will only apply to the DOCUMENT_TEXT_DETECTION feature.
ss...@gmail.com <ss...@gmail.com> #26
Page level orientation is definitely essential - and useful where the majority of the document rest upon one orientation (eg: letter, page, identity document).
Orientation of individual blocks, words etc would also be very useful if the information already exists through the model, and would be greatly desired.
Cheers,
Orientation of individual blocks, words etc would also be very useful if the information already exists through the model, and would be greatly desired.
Cheers,
ri...@gmail.com <ri...@gmail.com> #27
I think there should be an option field where you can decide if the orientation should also be given for individual blocks, words etc.
er...@gmail.com <er...@gmail.com> #28
Ideally we would have word-level information around rotation as well
mi...@gmail.com <mi...@gmail.com> #29
Great to hear that this feature will be added in the near future!
pe...@gmail.com <pe...@gmail.com> #30
+1 please!
ri...@gmail.com <ri...@gmail.com> #31
+1
to...@gmail.com <to...@gmail.com> #32
+1
sa...@gmail.com <sa...@gmail.com> #33
+1
mi...@gmail.com <mi...@gmail.com> #34
Any news on this feature yet? Will there be a beta version of the Cloud Vision API containing this? I'm really looking forward to this :)
ar...@gmail.com <ar...@gmail.com> #35
Any news?
za...@gmail.com <za...@gmail.com> #36
+1
[Deleted User] <[Deleted User]> #37
+1
sa...@gmail.com <sa...@gmail.com> #38
+1
ho...@gmail.com <ho...@gmail.com> #39
+1
ds...@gmail.com <ds...@gmail.com> #40
+1
tm...@gmail.com <tm...@gmail.com> #41
+1! Urgently needed.
sa...@gmail.com <sa...@gmail.com> #42
+1 We really need this one too.
ab...@gmail.com <ab...@gmail.com> #43
+1
th...@gmail.com <th...@gmail.com> #44
Orientation information should now be returned in 'boundingBox' object for Page Blocks [1], Paragraphs [2], Words [3], and Symbols [4] for all versions of the Vision API (v1, v1p3beta1, and v1p4beta1) [5].
[1]https://cloud.google.com/vision/docs/reference/rest/v1p4beta1/AnnotateImageResponse#block
[2]https://cloud.google.com/vision/docs/reference/rest/v1p4beta1/AnnotateImageResponse#paragraph
[3]https://cloud.google.com/vision/docs/reference/rest/v1p4beta1/AnnotateImageResponse#word
[4]https://cloud.google.com/vision/docs/reference/rest/v1p4beta1/AnnotateImageResponse#symbol
[5]https://cloud.google.com/vision/docs/reference/rest/
[1]
[2]
[3]
[4]
[5]
ib...@gmail.com <ib...@gmail.com> #45
Thanks for the update on this one, great to see it has been fixed.
However, could not see what additional information is provided to infer the orientation of an image in the updated boundingBox response?
What are the names of the elements in the response that provide this, and maybe an example of how to determine how an image should be operated on to make it upright?
Cheers!
However, could not see what additional information is provided to infer the orientation of an image in the updated boundingBox response?
What are the names of the elements in the response that provide this, and maybe an example of how to determine how an image should be operated on to make it upright?
Cheers!
ac...@gmail.com <ac...@gmail.com> #46
When orientation is detected, it is represented by the boundingBox vertex positions returned by the object types listed in comment#44 . Here is the documented description for the 'Page Block' type orientation:
"The vertices are in the order of top-left, top-right, bottom-right, bottom-left. When a rotation of the bounding box is detected the rotation is represented as around the top-left corner as defined when the text is read in the 'natural' orientation. For example:
when the text is horizontal it might look like:
0 ------- 1
| |
3 ------- 2
when it's rotated 180 degrees around the top-left corner it becomes:
2 ------- 3
| |
1 ------- 0
and the vertex order will still be (0, 1, 2, 3)."
"The vertices are in the order of top-left, top-right, bottom-right, bottom-left. When a rotation of the bounding box is detected the rotation is represented as around the top-left corner as defined when the text is read in the 'natural' orientation. For example:
when the text is horizontal it might look like:
0 ------- 1
| |
3 ------- 2
when it's rotated 180 degrees around the top-left corner it becomes:
2 ------- 3
| |
1 ------- 0
and the vertex order will still be (0, 1, 2, 3)."
rj...@gmail.com <rj...@gmail.com> #47
Can someone please add an example of where to find and how to extract this image-rotation?
gr...@bitvore.com <gr...@bitvore.com> #48
Well there be any support for this for OBJECT_DETECTION? I scan a lot of film negatives, and it would be amazing to guess the correct orientation of the image with a confidence score.
[Deleted User] <[Deleted User]> #49
Hello,
I don't understand
description: "0,13"
bounding_poly {
vertices {
x: 214
y: 147
}
vertices {
x: 215
y: 131
}
vertices {
x: 223
y: 132
}
vertices {
x: 222
y: 148
}
}
How should I interpret this to get the page orientation of my image (which is mostly text) ?
me...@writer.com <me...@writer.com> #50
+1
wa...@gmail.com <wa...@gmail.com> #51
+1
br...@writer.com <br...@writer.com> #52
[Deleted User] <[Deleted User]> #53
+1
ke...@gmail.com <ke...@gmail.com> #54
+1
gu...@gmail.com <gu...@gmail.com> #55
+1
sd...@neuropace.com <sd...@neuropace.com> #56
+1
[Deleted User] <[Deleted User]> #57
+1. Couldn't this be added in the last four and a half years this issue has been open?
li...@gmail.com <li...@gmail.com> #58
+1
[Deleted User] <[Deleted User]> #59
+1
mb...@yoobic.com <mb...@yoobic.com> #60
+1
mi...@gmail.com <mi...@gmail.com> #61
+1
zi...@gmail.com <zi...@gmail.com> #62
+1
rm...@google.com <rm...@google.com>
[Deleted User] <[Deleted User]> #63
+1
my...@gmail.com <my...@gmail.com> #64
+1
xi...@gmail.com <xi...@gmail.com> #65
+1
sh...@gmail.com <sh...@gmail.com> #66
+1
ni...@gmail.com <ni...@gmail.com> #67
+1
mo...@searce.com <mo...@searce.com> #68
Has this issue been solved yet?
My OCR results also seem to be output of images that are randomly rotated 0, 90, 180 or 270 degrees of original images.
My OCR results also seem to be output of images that are randomly rotated 0, 90, 180 or 270 degrees of original images.
am...@gmail.com <am...@gmail.com> #69
I see this is still in assigned status even after 4 years. Please expedite as it is already working in OCR, and needs to be exposed to the customers via response.
sa...@gmail.com <sa...@gmail.com> #70
I think I might have sorted it out, even if it is not perfect of course.
horizontal_text = []
for page in response.full_text_annotation.pages:
for block in page.blocks:
for paragraph in block.paragraphs:
prev_word = None
current_line = []
for word in paragraph.words:
word_text = ''.join([symbol.text for symbol in word.symbols])
y_diff = word.bounding_box.vertices[3].y - word.bounding_box.vertices[0].y
x_diff = word.bounding_box.vertices[3].x - word.bounding_box.vertices[0].x
if x_diff == 0:
angle = 90
else:
angle = abs(math.degrees(math.atan(y_diff / x_diff)))
if 45 <= angle < 135:
if prev_word and word.bounding_box.vertices[0].y - prev_word.bounding_box.vertices[3].y > \
threshold:
horizontal_text.append(''.join(current_line).strip())
current_line = []
current_line.append(' ' + word_text)
prev_word = word
if current_line:
horizontal_text.append(''.join(current_line).strip())
return '\n'.join(horizontal_text)
pi...@gmail.com <pi...@gmail.com> #71
Google is evil.
aa...@gmail.com <aa...@gmail.com> #72
+1 I created a bug about this today but have just realized this is talking about the same issue I am having, 6 years on and it's still not fixed!!
na...@lookdeep.health <na...@lookdeep.health> #73
+1
ti...@gmail.com <ti...@gmail.com> #74
+1
mi...@outfit7.com <mi...@outfit7.com> #75
+1
da...@fingerprint.de <da...@fingerprint.de> #76
+1
jw...@google.com <jw...@google.com> #77
+1 just had a retail select customer ask for this
Description
The beta of AutoML Vision is great but it would be VERY useful to be able to download the model / graph / weights in a number of common ML formats.
For me having an exported Tensorflow model / graph / weights would be great. It would be even better if there was an export option for the Movidius Neural Compute Stick.
I ask for this feature as my project works off a live feed from a camera inside a cat flap and needs to see if a cat has prey in its mouth. By the time a web request to the detection API came and returned, the cat would be in the house so downloading the model onto a Movidius NCS for offline detection would be an enormous help with the recognition speed.