$ voice2json --debug train-profile
ImportError: numpy.core.multiarray failed to import
Traceback (most recent call last):
File "/usr/lib/voice2json/.venv/lib/python3.7/site-packages/deepspeech/impl.py", line 14, in swig_import_helper
return importlib.import_module(mname)
File "/usr/lib/python3.7/importlib/__init__.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1006, in _gcd_import
File "<frozen importlib._bootstrap>", line 983, in _find_and_load
File "<frozen importlib._bootstrap>", line 967, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 670, in _load_unlocked
File "<frozen importlib._bootstrap>", line 583, in module_from_spec
File "<frozen importlib._bootstrap_external>", line 1043, in create_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
ImportError: numpy.core.multiarray failed to import
Building the v2.0 tag (or even master) using docker does not:
E: The repository 'http://security.ubuntu.com/ubuntu eoan-security Release' does not have a Release file.
And just bumping the image tag to ":groovy" caused subsequent silliness, so this project is obviously only for folks who enjoy fighting with build systems (and that matches my experience of anything in the world that touches Numpy and friends)
Good FLOSS speech recognition and TTS is badly needed. Such interaction should not be left to an oligoply with bad history of not respecting users freedoms and privacy.
Mozilla CommonVoice is definitely trying. I always do a few validations and a few clips if I have a few minutes to spare, and I recommend everyone does. They need volunteers to validate and upload speech clips to create a dataset.
I like the idea, and decided to try doing some validation. The first thing I noticed is that it asks me to make a yes-or-no judgment of whether the sentence was spoken "accurately", but nowhere on the site is it explained what "accurate" means, or how strict I should be.
(The first clip I got was spoken more or less correctly, but a couple of words are slurred together and the prosody is awkward. Without having a good idea of the standards and goals of the project, I have no idea whether including this clip would make the overall dataset better or worse. My gut feeling is that it's good for training recognition, and bad for training synthesis.)
This seems to me like a major issue, since it should take a relatively small amount of effort to write up a list of guidelines, and it would be hugely beneficial to establish those guidelines before asking a lot of volunteers to donate their time. I don't find it encouraging that this has been an open issue for four years, with apparently no action except a bunch of bikeshedding: https://github.com/common-voice/common-voice/issues/273
After listening to about 10 clips your point becomes abundantly clear.
One speaker, who sounded like they were from the mid-west United States, was dropping the S off words in a couple clips. I wasn't sure if it was misreads or some accent I'd never heard.
Another speaker, with a thick accent that sounded European, sounded out all the vowels in circuit. Had I not had the line being read, I don't think I'd have understood the word.
I heard a speaker with an Indian accent who added a preposition to the sentence that was inconsequential but incorrect none the less.
I hear these random prepositions added as flourishes frequently with some Indian coworkers, does anyone know the a reason? It's kind of like how American's interject "Umm..." or drop prepositions (e.g. "Are you done your meal?") and I almost didn't pick up on it. For that matter where did the American habit of dropping prepositions come from? It seems like it's people in the North East primarily.
I can't quite imagine superfluous prepositions (could you give an example?) but I have found it slightly amusing learning Hindi and coming across things where I think Oh! That's why you sometimes hear X from Indian English speakers, it's just a slightly 'too' literal¹ mapping from Hindi, or trying to use a grammatical construction that doesn't really exist in English, like 'topic marking'.
[¹] If that's even fair given it's a dialect in its own right - Americans also say things differently than I would as a 'Britisher'
That's not one I've heard. Examples that come to mind are 'even I' (which seems closer to 'I too' than the 'you'd scarcely believe it but I' that it naturally sounds to me), 'he himself' (or similar subject emphasis), and adverb repetition.
I'd say it's mostly subtler (I suppose that should be the expected distribution!) things I've noticed though, they're just harder to recall as a result.
(Just want to emphasise I'm not making fun of anybody or saying anything's wrong, in case it's not clear in text. I'm just enjoying learning Hindi, fairly interested in language generally, and interested/amused to notice these things.)
Just thought of another - '[something is] very less' - which comes, presumably, from कम being used for both little/few and less than.
Hindi is much more economical, to put it literally, one says things like 'than/from/compared to orange, lemon is sour', and 'orange is little/less [without comparison] sour'.
Which, I believe, is what gives rise to InE sentences like 'the salt in this is very less' (it needs more salt, there's very little).
I downloaded the (unofficial) Common Voice app [1] and it provides a link to some guidelines [2], which also aren't official but look sensible and seem like the best there is at the moment.
If you read the doc, it says voice2json is layer on top of the actual voice recognition engine. And it supports mozilla deep speech, pocket sphinx and a few others as the underlying engine.
I've used the deepspeech project a fair amount and it is good. It's not perfect, certainly, and it honestly isn't good enough yet for an accurate transcription in my mind, but it's good. Easy to work with, pretty good results, and all the right kinds of free.
That's fine for training your own model, but I don't think you could distribute the training set. That seems like a clear copyright violation, against one of the groups that cares most about copyright.
Maybe you could convince a couple of indie creators or state-run programs to licence their audio? But I'm not sure if negotiating that is more efficient than just recording a bit more audio, or promoting the project to get more volunteers.
It would likely be a lot easier for someone from within the BBC, CBC, PBS, or another public broadcaster to convince their employer to contribute to the models. These organizations often have accessibility mandates with real teeth and real costs implementing that mandate. The work of closed captioning, for example, can realistically be improved by excellent open source speech recognition and TTS models without handing all of the power over to Youtube and the like.
It would still be an uphill battle to convince them to hand over the training set but the legal department can likely be convinced if the data set they contribute back is heavily chopped up audio of the original content, especially if they have the originals before mixing. I imagine short audio files without any of the music, sound effects, or visual content are pretty much worthless as far as IP goes.
That's fine for training your own model, but I don't think you could distribute the training set. That seems like a clear copyright violation, against one of the groups that cares most about copyright.
I'm not sure that is a clear copyright violation. Sure, at a glance it seems like a derivative work, but it may be altered enough that it is not. I believe that collages, and reference guides like cliff notes are both legal.
I think a bigger problem would be that the scripts, and even the closed captioning, rarely match the recorded audio 100%
And also... it's not like the program actually contains a copy of the training data, right? The training data is a tool which is used to build a model.
How is it different from things like GPT3 which (unless I’m mistaken) is trained on a giant web scrape? I thought they didn’t release the model out of concerns for what people would do with a general prose generator rather than any copyright concerns?
Generally a ML model transforms the copyrighted material to the point where it isn't recognizable, so it should be treated as its own unrelated work that isn't infringing or derivative. But then you have e.g. GPT that is reproducing some (largeish) parts of the training set word-for-word, which might be infringing.
Also I don't think there have been any major court cases about this, so there's no clear precedent in either direction.
There are some that say that the Google Books court case is precedent for ML model stuff, if you search back through my comment history you will find links.
I am aware, I'm asking if the model, however, is infringing. Surely you can't distribute them in a dataset but is training on copyrighted data legal, and can you distribute that model?
All text written by a human in the US is automatically copyright the author. So if an engine trained on works under copyright is a derivative work, GPT3 and friends have serious problems.
I expect that wouldn't be perfect, though. Sometimes the cut that makes it into the final product doesn't exactly match the script. Sometimes it's due to an edit, other times it's due to an actor saying something similar to but not exactly what the script says, but the director deciding to just go with it.
What might work better is using closed captions or subtitles, but I've also seen enough cases where those don't exactly match the actual speech either.
Good speech recognition generally requites massive mountains of training data, both labelled and unlabelled.
Massive mountains of data tends to be incompatible with opensource projects. Even Mozilla collecting user statistics is pretty controversial. Imagine someone like Mozilla trying to collect hundreds of voice clips from each of tens of millions of users!!
> Really complicated question, but considering the free world got wikipedia and openstreetmaps, I'd bet we'll find a way.
Both of those involve entering data about external things. Asking people to share their own data is another thing entirely—I suspect most people, me included, are much more suspicious about that.
Then you need a lot of people that listen to those 12B hours of audio, and multiple listeners agree for each chunk of audio that what is spoken corresponds to the transcript.
Yes, but then you don't need Mozilla collecting read speech samples. You can just scrape any audio out there, run speech activity detection, and there you go.
Not an expert on any of this, but wouldn’t already published content (public or proprietary) such as Youtube videos, audiobooks, tv interviews, movies, tv programs, radio programs, podcasts, etc. be useful and exempt from privacy concerns?
Do user collected clips have soemthing so special to the point that it’s critical to collect them?
Another problem is that the models tend to get very very large for what I've seen. A gigabyte to 10s of gigabyes is an undesirable requirement on your local machine.
Not sure about others, but DeepSpeech also distributes a "lite" model that's much smaller and suitable for mobile devices. Not sure how its accuracy compares to the full model though.
It's well-documented and works basically out of box. I wish the STT models bundled were closer to the quality of Kaldi but the ease-of-use has no comparisons.
And maybe with time it will surpass Kaldi in quality too.
There are a bunch of good libraries that work offline out there for speech recognition -- CMUSphinx[0] has been around a long time and work seems to have shifted a little bit to Kaldi[1] and Vosk[2] (?). Julius is still going strong as well[3].
CMUSphinx and Julius have been around for ~10+ years at this point.
GNU is heavily skewed to developer tools and infrastructure, and gcc is no counterexample.
There are so many classes of software where this does not work. Pretty much anything for heavily regulated industries is not well served by FLOSS. There are few open source insurance software or medical records systems (the few that exist are highly niche and/or limited), EDA/CAD is not well served by FLOSS (I’ve toyed with FreeCAD, but even hobbyists gravitate to Fusion). Outside of developer tooling and infrastructure: commercial, generally closed source, closed development software is king.
* besides the hard part of standing up an EMR is not installing a prepackaged software.
There's more involved in than just raw CPU cycles. It's not something that is easily adapted to BOINC, but trying to offload things to BOINC to free up clusters better suited to training models might make sense.
Indeed, and it doesn't have to be as "machine learning" as the big ones.
A FLOSS system would only have my voice to recognise and I would be willing to spend some time training it. Very different usecase from a massive cloud that should recognise everyone's voice and accent.
pico2wav with the en-gb voice seems not too bad for TTS. I had reasonable luck in limited domain speech recognition with pocketsphinx, but it does need some custom vocabulary.
Granted, maybe this is "not good enough", but I feel like I got pretty far with pico2wave, pocketsphinx plus 1980's Zork level "comprehension" technology.
And the open source status of pico2wave is a bit questionable, I'll grant you that.
Author here. Thanks to everyone for checking out voice2json!
The TLDR of this project is: a unified command-line interface to different offline speech recognition projects, with the ability to train your own grammar/intent recognizer in one step.
My apologies for the broken packages; I'll get those fixed shortly. My focus lately has been on Rhasspy (https://github.com/rhasspy/rhasspy), which has a lot of the same ideas but a larger scope (full voice assistant).
Questions, comments, and suggestions are welcomed and appreciated!
I haven't seen this yet, but I imagine it would involve running at least "voice2json record-command | voice2json transcribe-wav | jq .text". This will record a single command (until silence), and output the text transcription.
I wonder if it would be possible to map vim keybindings to sounds and effectively drive the editor with the mouth when the hands are otherwise occupied. It might be possible to use sounds that compose into pronounceable words with minimal syllables for combinations. What would vim bindings look like as a concise command language suited to human vocalization?
E.g. maybe "dine" maps to d$ and "chine" to c$. So as in keyboard vim you can guess what "dend" and "chend" do.
I do this successfully for work using https://talonvoice.com/ - initial learning curve is steep, but once you learn how to configure and hack on the commands, you can be very effective. I use it maybe half the day to combat lingering RSI symptoms, and with some work I could probably use it for 98% of input for the computer. Some people do use it for 100% afaik
I wonder if this would pass the Debian Deep Learning Team's Machine Learning policy, which requires public data under a libre license that is retrainable using software under a libre license, without any proprietary drivers:
I think many of the available Kaldi/DeepSpeech models would pass, at least with the "Type-F Reproducibility". The pocketsphinx models would not, however, since they were trained on private datasets.
My aim has been to train "good enough" models for any public/free data I can get my hands on.
Really interesting use of intents and entities. I feel like some of this is reinventing the wheel, since there is already a grammar specification, but novel use of intents/entities. https://www.w3.org/TR/speech-grammar/
Yeah, in my experience no one uses or supports that specification, which is a shame because if you're using something like AWS Connect with AWS Lex for telephony IVR, you can't just create a grammar and then have AWS Lex figure out how to turn its recognized speech-to-text into something that matches a grammar rule. Thus, Lex will return speech-to-text results that are according to general English grammar rules, rather than what you might have prompted the user to reply with. You'll be unpleasantly surprised if you think that defining a custom entity as alphanumeric always prevents the utterance "[wʌn]" as sometimes matching "won" instead of "one" or "1".
Edit - Sorry, I realize that's a tangent. What I'm saying is that when I was evaluating speech to text engines for things like IVR systems using AWS and Google, neither of them supported SRGS. Microsoft does, I think, but they didn't have a telephony component, and IBM was ignored from the get go, so "no one" really means "two very large companies."
Some do, some don't, sure. Google STT for example supports class tokens natively. There are also services like uniMRCP that allow for certain SRGS grammar features to be used with Google STT, but they are limited in what constructs they support. I've worked pretty extensively with a platform called Verbio, and they fully support the SRGS grammar specification. I work in conversational AI, and when I do implementations, I have to evaluate the complexity of the use case and whether or not a full grammar will be needed and choose a STT provider based on that.
My templating language was inspired by JSGF, which seems to have informed the ABNF version of the W3C Speech Grammars. I don't support probabilities, though, since those are derived during the n-gram model generation.
I would have preferred to use a standard. Perhaps this is something for a future version.
I like rhasspy but the problem I have with it is that it's too much of a toolkit and less of an application. There's too many choices to pick for the different components.. I think they should pick one of each and really tune them so it works really well. This way they'd take a lot of complexity away from the user.
Agreed. I've at least added a "Recommended" option in the web UI that's language-specific.
Part of the problem is that language support varies dramatically between components. There's usually a pretty obvious "best" set for English, but it gets more difficult with other languages.
The goal of voice2json is to provide a common layer on top of existing open source engines. This common layer lets you train custom speech/intent models with having to know the details of each engine.
Excellent! I just installed MyCroft the other day to play around with it; while it looks like a great start, two odd things. The first is obvious, which is the online/offline thing.
The second was a little surprising (and maybe I missed it?) There was not much in the way of easily accessing transcribed output to and from shell scripts?
Not familiar with any of these tech, but would it be better to get the intent by
voice2txt command.wav | txt2intent
? Or the intent analyzation actually requires the sound data (what are the cases of the same phrase expressing different intent, or how do we even define / categorize intent in this context)
Would love to see a demo integrating this with an IDE for either voice to code or for voice commands to navigate menus. I think the killer application would layer voice and traditional input rather than replace.
I would really like this technology to takeoff in mobile apps. When interacting with my mobile phone often it is more convenient to do navigation by voice than by using my finger.
neat. would be even neater if it used state to provide a prior on likely intents. (ie. in its most simple form, if you know the light is on, "turn on the light" has a prior of 0)
I can finally build the Jarvis home assistant I dreamt of when first learning coding in high school. To bad now I know voice assistant widgets generally are useless.
Speech recognition is actually orthogonal to AI. In my day the AI prototypes (like ELIZA) were basically chat bots. Speech recognition is now very sophisticated and accurate. Determining meaning from human language (spoken or written) is far more advanced than it used to be but still kinda sucks.
Yes, text to speech has become impressively good and can basically be relied upon. But siri/alexa still really disappoint me and dont seem to be much better than a list of basic rules one could program up.
It's not quite clear, but do you need to sacrifice your privacy in any way to use it?
E.g. sending the data to some service in order to get trained model?
It's entirely unpackaged: https://repology.org/projects/?search=voice2json https://pkgs.org/search/?q=voice2json
Docker image is broken, how'd that happen?