User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
Admin
completely blind computer geek, lover of science fiction and fantasy (especially LitRPG). I work in accessibility, but my opinions are my own, not that of my employer. Fandoms: Harry Potter, Discworld, My Little Pony: Friendship is Magic, Buffy, Dead Like Me, Glee, and I'll read fanfic of pretty much anything that crosses over with one of those.
keyoxide: aspe:keyoxide.org:PFAQDLXSBNO7MZRNPUMWWKQ7TQ
Location
Ottawa
Birthday
1987-12-20
Pronouns
he/him (EN)
xmpp fastfinge@im.interfree.ca
keyoxide aspe:keyoxide.org:PFAQDLXSBNO7MZRNPUMWWKQ7TQ
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@cachondo @jscholes @FreakyFwoof @amir They don't have much choice. A lot of the libraries NVDA depends on are stopping 32-bit support this year.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@jscholes @cachondo @FreakyFwoof @amir My understanding is that when this comes to addons, it's going to require some kind of secure addons API/layer. And it won't be ready for 2026.1, or maybe not even 2026.2.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@FreakyFwoof @cachondo @amir You should be able to get either Gemini or Codex to help you, depending on what AI you have access to. The workflow would be:
1. download gemini-cli or codex-cli, and get them installed and configured.
2. clone all of the sourcecode from
github.com/fastfinge/eloquence_64/
3. Delete the tts.txt and tts.pdf files, so you don't confuse it with incorrect documentation.
4. Find any API documentation for orphius that's available, and add it into the folder.
4. Run codex-cli or gemini-cli, and tell it something like: "Using the information about how to develop NVDA addons you can find in agents.md, and the information about the Orphius API I've provided in the file Orphius-documentation-filename.txt, I would like you to modify the code in this folder to work with Orpheus instead of eloquence."

It will go away for five or ten minutes, ask you for permission to read and write the files it's interested in, and then give you something that mostly works. Now, build the addon, run it, and tell it about the errors and problems you have and ask it to fix them. In the case of errors, include the error right from the NVDA log, and for bugs and problems, tell it exactly what it's doing wrong, and exactly what you want it to do instead. Keep doing this until you wind up with a working addon.

Think of AI as a particularly stupid programmer, and you're the manager in charge of the project. You should be able to get this done without paying anyone.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@cachondo @amir I've heard from a second hand source that they are, yes. But I haven't verified that.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@pixelate @PepperTheVixen If you have a sample of someone talking while chewing gum, you can absolutely make that happen.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@pixelate @PepperTheVixen If you give chatterbox-tts an ASMR recording to clone, you can absolutely get it to make lip smacking noises.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@PepperTheVixen The reason it's grating is because unlike Eloquence and dectalk, Espeak only uses formant synthesis for the vowel sounds. For consonants and plosives, it instead uses concatenative recordings based on human speech. That's why even when you switch to a voice that sounds less sharp, the "t", "b", "p", and other sounds are still too sharp. This seems to be the primary cause of the fatigue most people experience while using ESpeak.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@svenja I vaguely remember writing something like that a while ago. The list I linked has all of the games I remember putting in that comment, though.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@Landon205 There's already addons that do that.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@svenja Was it this? gist.github.com/Molitvan/50e3b5060ab9465b1da895155d5c0480
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
The State of Modern AI Text To Speech Systems for Screen Reader Users: The past year has seen an explosion in new text to speech engines based on neural networks, large language models, and machine learning. But has any of this advancement offered anything to those using screen readers? stuff.interfree.ca/2026/01/05/ai-tts-for-screenreaders.html
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@pitermach Even Mona only indicates it after reading the text. IMHO, it should be before! I can see myself replying to a toot in a language the author doesn't speak.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
So one of the problems with automated based on is that it's in this weird, awkward place now where it's not good enough to seem native, but it is good enough not to be immediately obvious. I had forgotten I set my browser to automatically translate pages in several languages to English for me. Then I clicked on a , read the first couple chapters, and was getting really annoyed why one character kept changing Gender's, and the name of a location wasn't consistent from chapter to chapter. Turns out the author was writing in , I hadn't noticed the language tag, and when I clicked the browser just automatically translated without clearly telling me. Thank goodness I realized, just in time, before I posted a comment on her work pointing out the issues with writing quality and offering editing help! Yikes! That doesn't mean that I'm against Machine Translation or even think it's a bad thing. But as it gets better and better, and starts living in this awkward uncanny valley, we really need to carefully rethink user interfaces. Maybe disable all form inputs until the user clicks "I'm aware this was translated" somewhere? I had the comment all written, folks; I was that close to posting. Maybe put "translated from" in some obvious place like the title bar? I don't know what the answer is, but I do know that we need to be putting way more thought into it.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@jscholes If you long press on the lock screen, you can customize what shows up on there. I believe there's a way to turn off Siri suggestions. I use my watch for the time and date, so I left them on. But if I didn't have a watch that would be infuriating.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@silvermoon82 Also, if it's not a western, colonial European, or east-Asian language, screen reader support for it will be terrible. As far as I'm aware, there is zero support for any indigenous languages spoken by the people living in what are now Canada and the United States, in any screen reader or text to speech voice, anywhere. Same goes for many middle-eastern and African languages, as well as stuff like Irish Gaelic. If you speak those languages and are blind, your only real choice is...learn a different language and never use your phone or computer to read anything in your native language. Somehow, I doubt anyone on the fediverse is surprised by this.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@silvermoon82 It depends. If the user speaks both languages, and the lang attribute on the HTML has been set correctly, then the screen reader will read the correct text in the correct language. If the HTML lang attribute is not set, the user can change the language manually using a screen reader hotkey. However, these things assume the user has installed and configured voices that speak both languages. For example, I currently do not have a Russian voice installed (because I don't understand Russian anyway), so my screen reader will just read out unicode character codes, because it has no way of pronouncing the Russian alphabet. There are also addons that can recognize the language text is written in, and change the screen reader language automatically even if the language attribute isn't set, but most folks don't bother with these. As for translation: there are third party addons that can take all the text spoken by a screen reader, translate it into another language, and then speak it. However, this introduces lag, and sends everything the screen reader says off to a third party translation service. I would only use something like this if I was, for example, playing a Japanese game where I knew all text would be in Japanese, so my screen reader would just translate to English before speaking. I wouldn't just leave it on all the time.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@cachondo LOL it's the next dubstep hit! I think the blastbay voices might be our best bet. However, he's got to do something about the phonimizer. The voice quality is good, but it's based on the CMU dictionaries, so there are just way too many pronunciation issues.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@TomGrant91 @mckensie @amir I use enafore.social myself.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@TomGrant91 @mckensie @amir Right, but it's not missing. Tweesecake is stripping it out. It shows on every other client.
User avatar
🇨🇦Samuel Proulx🇨🇦 @fastfinge@interfree.ca
3mo
@TomGrant91 @mckensie @amir I can tell. Because only people using tweesecake have this issue. Even TWBlue doesn't.