Re : Re: [nvda-addons] Need help to vocalize a console
Hello Chris
I am responding a bit late... Hope this helps anyway.
Below is the chooseNVDAObjectOverlayClasses( function in my Matlab appModule. As written before, it allows to speak output, however, it speaks also each typed character whatever your keyboard echo is configured to speak.
def chooseNVDAObjectOverlayClasses(self, obj, clsList): import NVDAObjects.behaviors, NVDAObjects.window, NVDAObjects.window.winConsole if obj.windowClassName == "Edit" and obj.role == controlTypes.ROLE_EDITABLETEXT: #obj.STABILIZE_DELAY = 0 clsList[0:0] = [ NVDAObjects.behaviors.Terminal, NVDAObjects.window.DisplayModelLiveText, NVDAObjects.window.DisplayModelEditableText, ]
Cheers
Cyrille
----- Mail d'origine ----- De: Chris G. Pettyjohn (FTC) via Groups.Io <Chris.Pettyjohn@...> À: nvda-addons@nvda-addons.groups.io Envoyé: Tue, 09 Apr 2019 17:00:46 +0200 (CEST) Objet: Re: [nvda-addons] Need help to vocalize a console
hi Cyrille, Can you paste an example of how you added "NVDAObjects.behaviors.Terminal" and "NVDAObjects.window.DisplayModelLiveText" in chooseNVDAObjectOverlayClasses? I am having the same issue as you are with my consoles output. None of the new text is spoken unless i go and hunt for it. I created an appModule and have keystrokes that speak the last rows of text but I'd like it to speak as it comes on the screen. I use Micro Focus Reflection.
thanks, chris
|
|
Re: different pitch and speed for the keyboard then the general voice
Brian's Mail list account
So you want to find a way to increase response times and cut the time each character takes to say, but you do not want this in the rest of the screenreader? I don't actually see why, nor do I believe it can be done given the way the speech is handled by the synth. Obviously if you want some apps to have different settings that is one thing, but this will apply to input and output speaking.
Can you give a usage case for this effect? If you can hear the characters fast then set it to fast everywhere, if its merely read all that you want slow then use the profiles. Brian
bglists@blueyonder.co.uk Sent via blueyonder. Please address personal E-mail to:- briang1@blueyonder.co.uk, putting 'Brian Gaff' in the display name field.
toggle quoted messageShow quoted text
----- Original Message ----- From: "Isaac" <bigikemusic@gmail.com> To: <nvda-addons@nvda-addons.groups.io> Sent: Monday, April 15, 2019 10:21 AM Subject: Re: [nvda-addons] different pitch and speed for the keyboard then the general voice
Yes, when I type a key I want speed at 100 percent and pitch about at a 10 percent ----- Original Message ----- From: marcio via Groups.Io To: nvda-addons@nvda-addons.groups.io Sent: Monday, April 15, 2019 12:07 AM Subject: Re: [nvda-addons] different pitch and speed for the keyboard then the general voice
Hi, I didn't get your idea. Would you mind trying and explaining it again? Everything that deals with typing always catch my attention :)
Cheers, Marcio AKA Starboy
Follow or add me on Facebook
Em 15/04/2019 05:49, Isaac disse:
Hi, is there a way to set the typing keyboard to super fast. Thanks
|
|
Personally I dont use the copy to clipboard option directly, since I just include automatic descriptions If I think they are quite accurate or relevant for the situation, and I dont know this until I read the result. A different approach would be to add this at the bottom after result, maybe after a separator or blank lines. Even this could be optional and configurable.
Cheers
Enviado desde mi iPhone
toggle quoted messageShow quoted text
El 15 abr 2019, a las 7:21, Robert Hänggi <aarjay.robert@gmail.com> escribió:
On 14/04/2019, Noelia Ruiz <nrm1977@gmail.com> wrote: Hi, making another suggestion: I would like that the used plugin was mentioned at least in the title of browseable message or document result. This is useful when we copy some article with graphics and we want to include automatic descriptions with reference to the origin.
I don't like that as title, but you could append it to the clipboard result, something like: "Powered by OnlineImageDescriber". But you might need to mention the used engine as well.
Robert
Cheers
El 13/04/2019 a las 15:20, Larry Wang escribió: - a checkbox for OCR if short text should be spoken or always displayed. 'Use browse-able message for text result' is provided for this purpose. - Always Browse-able Message for image description but eventually replaced by a faked recognition object if any object positions are known. There is faked object result for image description. If you use the faked object in Azure analyzer position info is included. I set the faked position of general description to left right corner of the image. What comes after description is the name of celebrities landmarks and faces, location information are attached to object name.
Roughly 20% of ocr space result is timeout. May be 10 second timeout is still not enough for OCR?
As for GUI issue. May be I should put common options on top of engine settings for convenience.
On 2019/4/13 16:11, Robert Hänggi wrote: ized objects/faces but I don't see that here.
|
|
Re: different pitch and speed for the keyboard then the general voice
Yes, when I type a key I want speed at 100 percent
and pitch about at a 10 percent
toggle quoted messageShow quoted text
----- Original Message -----
Sent: Monday, April 15, 2019 12:07
AM
Subject: Re: [nvda-addons] different
pitch and speed for the keyboard then the general voice
Hi, I didn't get your idea. Would you mind trying and
explaining it again? Everything that deals with typing always catch my
attention :)
Em 15/04/2019 05:49, Isaac disse:
Hi, is there a way to set the typing keyboard to super fast.
Thanks
|
|
Re: different pitch and speed for the keyboard then the general voice
Hi,
I didn't get your idea. Would you mind trying and explaining it
again?
Everything that deals with typing always catch my attention :)
Em 15/04/2019 05:49, Isaac disse:
toggle quoted messageShow quoted text
Hi, is there a way to set the typing keyboard to super fast.
Thanks
|
|
Re : Re: Re : Re: [nvda-addons] Need help to vocalize a console
Hello Abdel
Sorry for late response. I have not had te to deal with Matlab console accessibility earlier.
I tried to add all or some of the classes you advised me. However, I cannot have the input rea word by word and the output read automatically in the same script. If I do not make any class overlay, I have input read word by word (according to my config), but output is not read.
If I add all these 3 classes Terminal,, NVDAObjects.window.DisplayModelLiveText, NVDAObjects.window.DisplayModelEditableText, I can have output read automatically, but input isread letter by letter whatever is my configuration. Moreover, I have the following error message when I start NVDA with the Matlab console focused :
Traceback (most recent call last): File "eventHandler.pyo", line 155, in executeEvent File "eventHandler.pyo", line 92, in __init__ File "eventHandler.pyo", line 100, in next File "NVDAObjects\IAccessible\__init__.pyo", line 1438, in event_caret File "NVDAObjects\window\edit.pyo", line 792, in event_caret File "documentBase.pyo", line 24, in makeTextInfo File "displayModel.pyo", line 264, in __init__ File "textInfos\offsets.pyo", line 407, in __init__ File "displayModel.pyo", line 586, in _getSelectionOffsets File "displayModel.pyo", line 541, in _getCaretOffset File "displayModel.pyo", line 172, in getCaretRect RuntimeError: displayModel_getCaretRect failed with res 1702
Any other idea and why I have this error that occurs only at NVDA startup ?
Thanks
Cheers
Cyrille
----- Mail d'origine ----- De: Abdel <abdelkrim.bensaid@...> À: nvda-addons@nvda-addons.groups.io Envoyé: Fri, 29 Mar 2019 19:09:06 +0100 (CET) Objet: Re: Re : Re: [nvda-addons] Need help to vocalize a console
Hi Cyrille,
I advise you to try in turn, the following overlay classes, obtained during the consultation of the log viewer after opening the Windows command prompt.
- NVDAObjects.Dynamic_IAccessibleWinConsoleWindowNVDAObject;
- NVDAObjects.window.winConsole.WinConsole;
- NVDAObjects.behaviors.Terminal;
- NVDAObjects.behaviors.LiveText.
You can exclude some classes if you see that they are not suitable.
Thanks.
Cheers,
Abdel.
Le 29/03/2019 à 13:37, Cyrille via Groups.Io a écrit :
Hello Robert
Thank you for your answer.
Yes I can activate system menu with Alt+Space or only Alt alone. This menu contains only the standard items (Restore, Move, Close, etc;)
I am using Matlab for team work. If I work alone on a project, I prefer using Python numpy/scipy.
Regards,
Cyrille
Cyrille
Hm, I work only with Octave and that's pretty accessible on the CLI.
Does the window have other elements?
Can you e.g. navigate to the system menu.
Robert
On 29/03/2019, Cyrille via Groups.Io <cyrille.bougot2@...> wrote:
> HelloI am trying to improve work with Matlab and NVDA on Windows. I would
> need help of someone having skills regarding NVDA console/terminal-like
> window support (no need to have Matlab on computer). Or someone who already
> tried something on Matlab.Today, Matlab UI is not accessible and its
> accessibility is not supported by Mathworks (Matlab's editor) for many
> years. Seems that it is due to version of Java Access Bridge imbedded into
> Matlab that do not match the one installed on the computer (nor version
> number, neither 32/64bit architecture). Anyway making Matlab accessible was
> not in their intention last time I contacted them.So I work only with Matlab
> console window (command line: "matlab -nodesktop" or "matlab -nojvm") and I
> concentrate only on improving accessibility of this console.Without any
> dedicated appModule, Matlab's console has the following properties:Developer
> info for navigator object:name: Nonerole: ROLE_EDITABLETEXTstates:
> STATE_MULTILINE, STATE_FOCUSABLE, STATE_FOCUSEDisFocusable: TruehasFocus:
> TruePython object: <NVDAObjects.Dynamic_IAccessibleEditWindowNVDAObject
> object at 0x0D28A2D0>Python class mro: (<class
> 'NVDAObjects.Dynamic_IAccessibleEditWindowNVDAObject'>, <class
> 'NVDAObjects.IAccessible.IAccessible'>, <class
> 'NVDAObjects.window.edit.Edit'>, <class
> 'NVDAObjects.behaviors.EditableTextWithAutoSelectDetection'>, <class
> 'NVDAObjects.behaviors.EditableText'>, <class 'editableText.EditableText'>,
> <class 'NVDAObjects.window.Window'>, <class 'NVDAObjects.NVDAObject'>,
> <class 'documentBase.TextContainerObject'>, <class
> 'baseObject.ScriptableObject'>, <class 'baseObject.AutoPropertyObject'>,
> <type 'object'>)description: Nonelocation: RectLTWH(left=282, top=164,
> width=1098, height=727)value: u'\xbb 'appModule: <'appModuleHandler'
> (appName u'matlab', process ID 16708) at address
> d1f4dd0>appModule.productName: u'MATLAB'appModule.productVersion:
> u'9.5.0.933748'TextInfo: <class
> 'NVDAObjects.window.edit.EditTextInfo'>windowHandle: 852420windowClassName:
> u'Edit'windowControlID: 99windowStyle: 1345323204windowThreadID:
> 13424windowText: u'\xbb 'displayText: u'\xbb For online documentation, see
> https://www.mathworks.com/supportFor product information, visit
> www.mathworks.com.\n[[[CONTENT_DELETED_MANUALLY]]]'
> (truncated)IAccessibleObject: <POINTER(IAccessible) ptr=0x2652148 at
> 7320a80>IAccessibleChildID: 0IAccessible event parameters:
> windowHandle=852420, objectID=-4, childID=0IAccessible accName:
> NoneIAccessible accRole: ROLE_SYSTEM_TEXTIAccessible accState:
> STATE_SYSTEM_FOCUSED, STATE_SYSTEM_FOCUSABLE, STATE_SYSTEM_VALID
> (1048580)IAccessible accDescription: NoneIAccessible accValue: u'\xbb 'When
> I type in the console, the text is announced (by word according to my
> configuration). However when I hit enter key, the output text is not spoken
> and I need to review it with review cursor.I would like to have output
> spoken.By adding "NVDAObjects.behaviors.Terminal" and
> "NVDAObjects.window.DisplayModelLiveText" in chooseNVDAObjectOverlayClasses,
> I succeed in having output spoken. However, in this case, whenn I type text,
> the text is spoken character by character and not by word anymore; my
> configuration to speak input by word is not honored anymore.Anybody could
> help me to match these 2 requirements :- have console output spoken- honor
> input configuration as wordAny idea is welcome since I do not have a good
> comprehension on how NVDA manages text in console.Thanks in
> advance.Regards,Cyrille
>
>
>
>
|
|
different pitch and speed for the keyboard then the general voice
Hi, is there a way to set the typing keyboard to
super fast.
Thanks
|
|
Also thanks to Jose Manuel delicado for telling me about this message as I missed it when it was posted.
Thanks.
|
|
Hi. Just wanted to come here and say that it's no problem to use word count as the name for the addon. Text unit counter sounds too complicated for the general user to understand. If you are not an NVDA developer, you probably don't know what a text unit stands for.
Cheers.
|
|
On 14/04/2019, Noelia Ruiz <nrm1977@gmail.com> wrote: Hi, making another suggestion: I would like that the used plugin was mentioned at least in the title of browseable message or document result. This is useful when we copy some article with graphics and we want to include automatic descriptions with reference to the origin.
I don't like that as title, but you could append it to the clipboard result, something like: "Powered by OnlineImageDescriber". But you might need to mention the used engine as well. Robert Cheers
El 13/04/2019 a las 15:20, Larry Wang escribió:
- a checkbox for OCR if short text should be spoken or always displayed. 'Use browse-able message for text result' is provided for this purpose. - Always Browse-able Message for image description but eventually replaced by a faked recognition object if any object positions are known. There is faked object result for image description. If you use the faked object in Azure analyzer position info is included. I set the faked position of general description to left right corner of the image. What comes after description is the name of celebrities landmarks and faces, location information are attached to object name.
Roughly 20% of ocr space result is timeout. May be 10 second timeout is still not enough for OCR?
As for GUI issue. May be I should put common options on top of engine settings for convenience.
On 2019/4/13 16:11, Robert Hänggi wrote:
ized objects/faces but I don't see that here.
|
|
Hello, Silly question, perhaps, but I remember seeing a while back that one could enable automatic reading of text as it changed on the screen? I don't see this in the list of hotkeys for the add on so am I missing something? Also - may I get the link where this can be downloaded again? Thanks!
toggle quoted messageShow quoted text
-----Original Message----- From: nvda-addons@nvda-addons.groups.io <nvda-addons@nvda-addons.groups.io> On Behalf Of Noelia Ruiz Sent: April 14, 2019 3:08 PM To: nvda-addons@nvda-addons.groups.io Subject: Re: [nvda-addons] Online OCR addon #addonrequestreview Hi, making another suggestion: I would like that the used plugin was mentioned at least in the title of browseable message or document result. This is useful when we copy some article with graphics and we want to include automatic descriptions with reference to the origin. Cheers El 13/04/2019 a las 15:20, Larry Wang escribió: - a checkbox for OCR if short text should be spoken or always displayed. 'Use browse-able message for text result' is provided for this purpose. - Always Browse-able Message for image description but eventually replaced by a faked recognition object if any object positions are known. There is faked object result for image description. If you use the faked object in Azure analyzer position info is included. I set the faked position of general description to left right corner of the image. What comes after description is the name of celebrities landmarks and faces, location information are attached to object name.
Roughly 20% of ocr space result is timeout. May be 10 second timeout is still not enough for OCR?
As for GUI issue. May be I should put common options on top of engine settings for convenience.
On 2019/4/13 16:11, Robert Hänggi wrote:
ized objects/faces but I don't see that here.
|
|
Hi, making another suggestion: I would like that the used plugin was mentioned at least in the title of browseable message or document result. This is useful when we copy some article with graphics and we want to include automatic descriptions with reference to the origin.
Cheers
toggle quoted messageShow quoted text
El 13/04/2019 a las 15:20, Larry Wang escribió: - a checkbox for OCR if short text should be spoken or always displayed. 'Use browse-able message for text result' is provided for this purpose. - Always Browse-able Message for image description but eventually replaced by a faked recognition object if any object positions are known. There is faked object result for image description. If you use the faked object in Azure analyzer position info is included. I set the faked position of general description to left right corner of the image. What comes after description is the name of celebrities landmarks and faces, location information are attached to object name. Roughly 20% of ocr space result is timeout. May be 10 second timeout is still not enough for OCR? As for GUI issue. May be I should put common options on top of engine settings for convenience. On 2019/4/13 16:11, Robert Hänggi wrote:
ized objects/faces but I don't see that here.
|
|

Larry Wang
- a checkbox for OCR if short text should be spoken or always displayed. 'Use browse-able message for text result' is provided for this purpose. - Always Browse-able Message for image description but eventually replaced by a faked recognition object if any object positions are known. There is faked object result for image description. If you use the faked object in Azure analyzer position info is included. I set the faked position of general description to left right corner of the image. What comes after description is the name of celebrities landmarks and faces, location information are attached to object name.
Roughly 20% of ocr space result is timeout. May be 10 second timeout is still not enough for OCR?
As for GUI issue. May be I should put common options on top of engine settings for convenience.
toggle quoted messageShow quoted text
On 2019/4/13 16:11, Robert Hänggi wrote: ized objects/faces but I don't see that here.
|
|

Larry Wang
The use of ui.message comes from the request of a person who uses OCR for gaming. He wanted to hear what is on screen then perform certain actions. He said dismissing fake object everytime is a bit inconveient when using Windows 10 OCR. According to recent feedback. Pop something up is a more common use case. May be I should make pop up window as default?
toggle quoted messageShow quoted text
On 2019/4/13 12:23, Noelia Ruiz wrote: Hi, I agree with most of this. I think that Machine learning could work sometimes, but many tomes doesn't work here.
5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. Maybe better to use browseable message, since when using Win20 OCR style, the result is also shown in browse mode, that is, both browseable message and OCR result are shown in different browse mode windows. Personally, I would remove the possibility of using just ui.message, since generally this is not confortable for braille with long messages like that.
Cheers
El 12/04/2019 a las 21:22, Andre Fisher escribió:
Hi.
A few comments and interesting observations.
1. On my end, the Machine Learning Engine by Oliver Edholm option doesn't work at all. This is also bothersome, as it is the default option. 2. The acronym 'API' should always bee in uppercase, so the option 'Use public api quota' is incorrect. 3. Change 'Max' to 'Maximum' in the number of candidates to avoid confusion. 4. Change option from 'Copy result to clipboard after recognition' to 'Copy recognition result to the clipboard' 5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. 6. When the above option is unchecked, and the mode used to show the result is similar to the one using Windows 10 OCR, navigation by line, word and paragraph is impossible. 7. The message shown when the option is checked has no title, but appears as 'NVDA message' 8. The option 'Swap the effect of repeated gesture with none repeated ones' isn't clear by it's name. I suggest removing it (see next point why). 9. There is a bug with the add-on. While copying to the clipboard works, there seems to be an issue with 'Swap the effect of repeated gesture with none repeated ones'. (option B) and 'Use browseable message for text result' (option A). When Option A is checked, and Option B is both checked and unchecked, a browsable message does appear. When option A is unchecked, and option B is both checked unchecked, the message appears in a Virtual document. So, option B seems useless, as even with it unchecked, the result still shows in a browse mode/virtual document. 10. 'Enable more verbose logging for debug purpose' should be '...purposes'
|
|

Larry Wang
Hi Andre,
Oliver's engine limit picture size to 10KB unlike other engine's
1MB to 4MB limit. If raw image is bigger than limit the addon will
resize the image. That is why Oliver's engine does not work
properly. I have reported this to him.
Feedback on labels is excellent, I will correct them.
As for the two options. There may be some misunderstanding. Maybe
I should write more about them in addon document.
The behaviour you described is desired.
Option A and B are irrelevant.
Option A controls which method to use when location information is
not needed (aka text result).
Default is ui.message is used. Result will be shown as one time
speech or braille message if A is checked ui.browseableMessage is
used. Result will be represented in a pop up window.
Option B controls which type of result is used when pressing
gesture different times.
By default, If you press once text result is used.
If you press twice a faked object like Windows 10 OCR is used.
If B is checked these behaviour is swapped.
In your feedback, when A is checked.
The browseable messages shown are actually different.
When B is checked a faked object is shown.
Note that it is faked, there is no real window or control on
screen. There is no visual feedback. The reason why NVDA does not
popup something is that such an object usually contains location
information of recognized text. You can use it for further
opertaion such as clicking on an inaccessible button. Pop up
something may break further opertaion.
If B is not checked. Result is shown in a pop up window with the
title 'NVDA message'. There is no location information in this
result.
toggle quoted messageShow quoted text
On 2019/4/13 3:22, Andre Fisher wrote:
Hi.
A few comments and interesting observations.
1. On my end, the Machine Learning Engine by Oliver Edholm option
doesn't work at all. This is also bothersome, as it is the default
option.
2. The acronym 'API' should always bee in uppercase, so the option
'Use public api quota' is incorrect.
3. Change 'Max' to 'Maximum' in the number of candidates to avoid
confusion.
4. Change option from 'Copy result to clipboard after recognition'
to 'Copy recognition result to the clipboard'
5. 'Use browseable message for text result' can be changed to
'Show recognition result in browse mode'.
6. When the above option is unchecked, and the mode used to show
the result is similar to the one using Windows 10 OCR, navigation
by line, word and paragraph is impossible.
7. The message shown when the option is checked has no title, but
appears as 'NVDA message'
8. The option 'Swap the effect of repeated gesture with none
repeated ones' isn't clear by it's name. I suggest removing it
(see next point why).
9. There is a bug with the add-on. While copying to the clipboard
works, there seems to be an issue with 'Swap the effect of
repeated gesture with none repeated ones'. (option B) and 'Use
browseable message for text result' (option A).
When Option A is checked, and Option B is both checked and
unchecked, a browsable message does appear.
When option A is unchecked, and option B is both checked
unchecked, the message appears in a Virtual document.
So, option B seems useless, as even with it unchecked, the result
still shows in a browse mode/virtual document.
10. 'Enable more verbose logging for debug purpose' should be
'...purposes'
|
|
Hi, a combo box would be also good. I don't know why readonly and buttons are used for braille and synthesizers, maybe for visual presentation, and really I don't know how visual controls are presented on the screen. About profiles, for simplicity, I would make all options configurable, so that people don't have to know what of them are just for a plugin and what general.
Cheers
toggle quoted messageShow quoted text
El 13/04/2019 a las 12:10, Robert Hänggi escribió: On 13/04/2019, Noelia Ruiz <nrm1977@gmail.com> wrote:
Robert's approach sounds good, at least regarding messages About the dialog, for me both approaches, the current one and Robert suggestions, are acceptable, since API keys deppend on each plugin, so maybe someone uses an own api key for just the most frequently used plugin, but if it needs to be changed, the api key would need to be modified too, so, at least in some cases, api key option would be as frequent like plugins choices. Personally, I would build the dialog as follows: - A readonly for OCR with the changer button. - The readonly for Image describer and the changer button. Why can't this be a combo box? The "Choose" or rather "Details" button would then open the settings for that particular engine. One important point is the scope: What should be changeable per profile and what settings should be general. I think the access methods could be global (i.e. what is in the dialog popping up), what do you think? On the other hand, scaling settings, the engine engaged and the detail selection could be per profile and should thus stay in the main panel (makes it necessary to change the content dynamically, depending on the engine selected). By the way, I would be glad if the Windows OCR could be selectable as well, it could profit from extended image filtering, e.g. scaling with sharpening, converting to grey scale, etc. This might yield better results. Robert
- This two main buttons would open dialogs with combo box to select, respectively, the OCR with their options, including api key for each one indepently, and the same for image describers. This would have the advantage that, if other plugins are added in future, options could be easily customized for each one.
Cheers
El 13/04/2019 a las 10:11, Robert Hänggi escribió:
On 13/04/2019, Noelia Ruiz <nrm1977@gmail.com> wrote:
Hi, I agree with most of this. I think that Machine learning could work sometimes, but many tomes doesn't work here.
5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. Maybe better to use browseable message, since when using Win20 OCR style, the result is also shown in browse mode, that is, both browseable message and OCR result are shown in different browse mode windows. Personally, I would remove the possibility of using just ui.message, since generally this is not confortable for braille with long messages like that. How about limiting ui.message output to one-liners and only for OCR?
I wouldn't use browseable message either. The recognition object would be fine if you replace the Windows line breaks (\CR\LF) with normal ones. It might be necessary to tweak the text infos because there are no positions for a description result, in other words, you would have to fake them. The reason behind using the recognition object is that at some point the recognized objects and faces might have a position assigned to them. If it was a button (e.g. a compact disc), it could perform an action. You can't achieve that with a browseable message (Sorry for the typo but I use browseable as it is used by NVDA). By the way, it is documented in the add-on help that Azure can deliver the approx position of recognized objects/faces but I don't see that here.
In summary: - a checkbox for OCR if short text should be spoken or always displayed. - Always Browseable Message for image description but eventually replaced by a faked recognition object if any object positions are known.
Oliver's engine works here but OCR Space gives me 99 % of the time a time out.
Also, I would turn the options in the GUI around such that the ones that are frequently changed are on the main panel. That is, the engine should be a combo box, not a read-only field whereas the API access details should be hidden in a sub-menu (button next to the combo) since they should work once they are set up.
Cheers Robert
Cheers
El 12/04/2019 a las 21:22, Andre Fisher escribió:
Hi.
A few comments and interesting observations.
1. On my end, the Machine Learning Engine by Oliver Edholm option doesn't work at all. This is also bothersome, as it is the default option. 2. The acronym 'API' should always bee in uppercase, so the option 'Use public api quota' is incorrect. 3. Change 'Max' to 'Maximum' in the number of candidates to avoid confusion. 4. Change option from 'Copy result to clipboard after recognition' to 'Copy recognition result to the clipboard' 5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. 6. When the above option is unchecked, and the mode used to show the result is similar to the one using Windows 10 OCR, navigation by line, word and paragraph is impossible. 7. The message shown when the option is checked has no title, but appears as 'NVDA message' 8. The option 'Swap the effect of repeated gesture with none repeated ones' isn't clear by it's name. I suggest removing it (see next point why). 9. There is a bug with the add-on. While copying to the clipboard works, there seems to be an issue with 'Swap the effect of repeated gesture with none repeated ones'. (option B) and 'Use browseable message for text result' (option A). When Option A is checked, and Option B is both checked and unchecked, a browsable message does appear. When option A is unchecked, and option B is both checked unchecked, the message appears in a Virtual document. So, option B seems useless, as even with it unchecked, the result still shows in a browse mode/virtual document. 10. 'Enable more verbose logging for debug purpose' should be '...purposes'
|
|
On 13/04/2019, Noelia Ruiz <nrm1977@gmail.com> wrote: Robert's approach sounds good, at least regarding messages About the dialog, for me both approaches, the current one and Robert suggestions, are acceptable, since API keys deppend on each plugin, so maybe someone uses an own api key for just the most frequently used plugin, but if it needs to be changed, the api key would need to be modified too, so, at least in some cases, api key option would be as frequent like plugins choices. Personally, I would build the dialog as follows: - A readonly for OCR with the changer button. - The readonly for Image describer and the changer button. Why can't this be a combo box? The "Choose" or rather "Details" button would then open the settings for that particular engine. One important point is the scope: What should be changeable per profile and what settings should be general. I think the access methods could be global (i.e. what is in the dialog popping up), what do you think? On the other hand, scaling settings, the engine engaged and the detail selection could be per profile and should thus stay in the main panel (makes it necessary to change the content dynamically, depending on the engine selected). By the way, I would be glad if the Windows OCR could be selectable as well, it could profit from extended image filtering, e.g. scaling with sharpening, converting to grey scale, etc. This might yield better results. Robert - This two main buttons would open dialogs with combo box to select, respectively, the OCR with their options, including api key for each one indepently, and the same for image describers. This would have the advantage that, if other plugins are added in future, options could be easily customized for each one.
Cheers
El 13/04/2019 a las 10:11, Robert Hänggi escribió:
On 13/04/2019, Noelia Ruiz <nrm1977@gmail.com> wrote:
Hi, I agree with most of this. I think that Machine learning could work sometimes, but many tomes doesn't work here.
5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. Maybe better to use browseable message, since when using Win20 OCR style, the result is also shown in browse mode, that is, both browseable message and OCR result are shown in different browse mode windows. Personally, I would remove the possibility of using just ui.message, since generally this is not confortable for braille with long messages like that. How about limiting ui.message output to one-liners and only for OCR?
I wouldn't use browseable message either. The recognition object would be fine if you replace the Windows line breaks (\CR\LF) with normal ones. It might be necessary to tweak the text infos because there are no positions for a description result, in other words, you would have to fake them. The reason behind using the recognition object is that at some point the recognized objects and faces might have a position assigned to them. If it was a button (e.g. a compact disc), it could perform an action. You can't achieve that with a browseable message (Sorry for the typo but I use browseable as it is used by NVDA). By the way, it is documented in the add-on help that Azure can deliver the approx position of recognized objects/faces but I don't see that here.
In summary: - a checkbox for OCR if short text should be spoken or always displayed. - Always Browseable Message for image description but eventually replaced by a faked recognition object if any object positions are known.
Oliver's engine works here but OCR Space gives me 99 % of the time a time out.
Also, I would turn the options in the GUI around such that the ones that are frequently changed are on the main panel. That is, the engine should be a combo box, not a read-only field whereas the API access details should be hidden in a sub-menu (button next to the combo) since they should work once they are set up.
Cheers Robert
Cheers
El 12/04/2019 a las 21:22, Andre Fisher escribió:
Hi.
A few comments and interesting observations.
1. On my end, the Machine Learning Engine by Oliver Edholm option doesn't work at all. This is also bothersome, as it is the default option. 2. The acronym 'API' should always bee in uppercase, so the option 'Use public api quota' is incorrect. 3. Change 'Max' to 'Maximum' in the number of candidates to avoid confusion. 4. Change option from 'Copy result to clipboard after recognition' to 'Copy recognition result to the clipboard' 5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. 6. When the above option is unchecked, and the mode used to show the result is similar to the one using Windows 10 OCR, navigation by line, word and paragraph is impossible. 7. The message shown when the option is checked has no title, but appears as 'NVDA message' 8. The option 'Swap the effect of repeated gesture with none repeated ones' isn't clear by it's name. I suggest removing it (see next point why). 9. There is a bug with the add-on. While copying to the clipboard works, there seems to be an issue with 'Swap the effect of repeated gesture with none repeated ones'. (option B) and 'Use browseable message for text result' (option A). When Option A is checked, and Option B is both checked and unchecked, a browsable message does appear. When option A is unchecked, and option B is both checked unchecked, the message appears in a Virtual document. So, option B seems useless, as even with it unchecked, the result still shows in a browse mode/virtual document. 10. 'Enable more verbose logging for debug purpose' should be '...purposes'
|
|
Also, I'm thinking that scripts with unassigned gestures could be included, like the synthesizer and Voice dialog, to open the main dialog and the two dialogs corresponding to OCR and Image describers, just some thoughts.
Cheers
toggle quoted messageShow quoted text
El 13/04/2019 a las 10:59, Noelia Ruiz via Groups.Io escribió: Robert's approach sounds good, at least regarding messages About the dialog, for me both approaches, the current one and Robert suggestions, are acceptable, since API keys deppend on each plugin, so maybe someone uses an own api key for just the most frequently used plugin, but if it needs to be changed, the api key would need to be modified too, so, at least in some cases, api key option would be as frequent like plugins choices. Personally, I would build the dialog as follows: - A readonly for OCR with the changer button. - The readonly for Image describer and the changer button. - This two main buttons would open dialogs with combo box to select, respectively, the OCR with their options, including api key for each one indepently, and the same for image describers. This would have the advantage that, if other plugins are added in future, options could be easily customized for each one. Cheers El 13/04/2019 a las 10:11, Robert Hänggi escribió:
On 13/04/2019, Noelia Ruiz <nrm1977@gmail.com> wrote:
Hi, I agree with most of this. I think that Machine learning could work sometimes, but many tomes doesn't work here.
5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. Maybe better to use browseable message, since when using Win20 OCR style, the result is also shown in browse mode, that is, both browseable message and OCR result are shown in different browse mode windows. Personally, I would remove the possibility of using just ui.message, since generally this is not confortable for braille with long messages like that. How about limiting ui.message output to one-liners and only for OCR?
I wouldn't use browseable message either. The recognition object would be fine if you replace the Windows line breaks (\CR\LF) with normal ones. It might be necessary to tweak the text infos because there are no positions for a description result, in other words, you would have to fake them. The reason behind using the recognition object is that at some point the recognized objects and faces might have a position assigned to them. If it was a button (e.g. a compact disc), it could perform an action. You can't achieve that with a browseable message (Sorry for the typo but I use browseable as it is used by NVDA). By the way, it is documented in the add-on help that Azure can deliver the approx position of recognized objects/faces but I don't see that here.
In summary: - a checkbox for OCR if short text should be spoken or always displayed. - Always Browseable Message for image description but eventually replaced by a faked recognition object if any object positions are known.
Oliver's engine works here but OCR Space gives me 99 % of the time a time out.
Also, I would turn the options in the GUI around such that the ones that are frequently changed are on the main panel. That is, the engine should be a combo box, not a read-only field whereas the API access details should be hidden in a sub-menu (button next to the combo) since they should work once they are set up.
Cheers Robert
Cheers
El 12/04/2019 a las 21:22, Andre Fisher escribió:
Hi.
A few comments and interesting observations.
1. On my end, the Machine Learning Engine by Oliver Edholm option doesn't work at all. This is also bothersome, as it is the default option. 2. The acronym 'API' should always bee in uppercase, so the option 'Use public api quota' is incorrect. 3. Change 'Max' to 'Maximum' in the number of candidates to avoid confusion. 4. Change option from 'Copy result to clipboard after recognition' to 'Copy recognition result to the clipboard' 5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. 6. When the above option is unchecked, and the mode used to show the result is similar to the one using Windows 10 OCR, navigation by line, word and paragraph is impossible. 7. The message shown when the option is checked has no title, but appears as 'NVDA message' 8. The option 'Swap the effect of repeated gesture with none repeated ones' isn't clear by it's name. I suggest removing it (see next point why). 9. There is a bug with the add-on. While copying to the clipboard works, there seems to be an issue with 'Swap the effect of repeated gesture with none repeated ones'. (option B) and 'Use browseable message for text result' (option A). When Option A is checked, and Option B is both checked and unchecked, a browsable message does appear. When option A is unchecked, and option B is both checked unchecked, the message appears in a Virtual document. So, option B seems useless, as even with it unchecked, the result still shows in a browse mode/virtual document. 10. 'Enable more verbose logging for debug purpose' should be '...purposes'
|
|
Re: Required Weather plus
Thank you so much friend.for your reply.
toggle quoted messageShow quoted text
On 4/13/19, Rémy Ruiz <remyruiz@gmail.com> wrote: Hy Nomi and all,
This version is provisional, until the new version comes out, currently the author is working on his code, correcting errors, etc.
Weather Plus (Apixu) version 5.5 using Adriano's Key):
http://www.nvda.it/files/plugin/weather_plus5.5.nvda-addon
Thanks for using the Weather_Plus add-on!
Cheers.
Rémy
Le 13/04/2019 à 08:46, Nomi a écrit :
Hi dear listers, Hope you'll all be fine. Friends I need an addon called weather plus, Actually I searched it on google, but Provided link was not working and giving error. So I request if anyone from the list has this add-on, Please Share it in the group. Regards
|
|
Robert's approach sounds good, at least regarding messages About the dialog, for me both approaches, the current one and Robert suggestions, are acceptable, since API keys deppend on each plugin, so maybe someone uses an own api key for just the most frequently used plugin, but if it needs to be changed, the api key would need to be modified too, so, at least in some cases, api key option would be as frequent like plugins choices. Personally, I would build the dialog as follows: - A readonly for OCR with the changer button. - The readonly for Image describer and the changer button. - This two main buttons would open dialogs with combo box to select, respectively, the OCR with their options, including api key for each one indepently, and the same for image describers. This would have the advantage that, if other plugins are added in future, options could be easily customized for each one.
Cheers
toggle quoted messageShow quoted text
El 13/04/2019 a las 10:11, Robert Hänggi escribió: On 13/04/2019, Noelia Ruiz <nrm1977@gmail.com> wrote:
Hi, I agree with most of this. I think that Machine learning could work sometimes, but many tomes doesn't work here.
5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. Maybe better to use browseable message, since when using Win20 OCR style, the result is also shown in browse mode, that is, both browseable message and OCR result are shown in different browse mode windows. Personally, I would remove the possibility of using just ui.message, since generally this is not confortable for braille with long messages like that. How about limiting ui.message output to one-liners and only for OCR? I wouldn't use browseable message either. The recognition object would be fine if you replace the Windows line breaks (\CR\LF) with normal ones. It might be necessary to tweak the text infos because there are no positions for a description result, in other words, you would have to fake them. The reason behind using the recognition object is that at some point the recognized objects and faces might have a position assigned to them. If it was a button (e.g. a compact disc), it could perform an action. You can't achieve that with a browseable message (Sorry for the typo but I use browseable as it is used by NVDA). By the way, it is documented in the add-on help that Azure can deliver the approx position of recognized objects/faces but I don't see that here. In summary: - a checkbox for OCR if short text should be spoken or always displayed. - Always Browseable Message for image description but eventually replaced by a faked recognition object if any object positions are known. Oliver's engine works here but OCR Space gives me 99 % of the time a time out. Also, I would turn the options in the GUI around such that the ones that are frequently changed are on the main panel. That is, the engine should be a combo box, not a read-only field whereas the API access details should be hidden in a sub-menu (button next to the combo) since they should work once they are set up. Cheers Robert
Cheers
El 12/04/2019 a las 21:22, Andre Fisher escribió:
Hi.
A few comments and interesting observations.
1. On my end, the Machine Learning Engine by Oliver Edholm option doesn't work at all. This is also bothersome, as it is the default option. 2. The acronym 'API' should always bee in uppercase, so the option 'Use public api quota' is incorrect. 3. Change 'Max' to 'Maximum' in the number of candidates to avoid confusion. 4. Change option from 'Copy result to clipboard after recognition' to 'Copy recognition result to the clipboard' 5. 'Use browseable message for text result' can be changed to 'Show recognition result in browse mode'. 6. When the above option is unchecked, and the mode used to show the result is similar to the one using Windows 10 OCR, navigation by line, word and paragraph is impossible. 7. The message shown when the option is checked has no title, but appears as 'NVDA message' 8. The option 'Swap the effect of repeated gesture with none repeated ones' isn't clear by it's name. I suggest removing it (see next point why). 9. There is a bug with the add-on. While copying to the clipboard works, there seems to be an issue with 'Swap the effect of repeated gesture with none repeated ones'. (option B) and 'Use browseable message for text result' (option A). When Option A is checked, and Option B is both checked and unchecked, a browsable message does appear. When option A is unchecked, and option B is both checked unchecked, the message appears in a Virtual document. So, option B seems useless, as even with it unchecked, the result still shows in a browse mode/virtual document. 10. 'Enable more verbose logging for debug purpose' should be '...purposes'
|
|