onilbent nirvelee kaelyn

0 views
Skip to first unread message

Gordon Neal

unread,
Aug 2, 2024, 11:11:04 PM8/2/24
to luangpardlletmysq

Improve your Greek Listening and reading comprehension with our Greek Audio texts. You can also find the Greek written texts, as well as the English translations on the www.omilo.com website.If you are at the Advanced level, and you are interested to learn more Greek, with Audio, as well as extra grammar and vocabulary exercises, for the advanced levels, then the Omilo Greek Advanced Workbooks are the perfect eBooks to have : Info at -greek-workbooks/

Hi all, please bear with me since im completely new to this, doing baby steps to understand how this works.
So, an audio(wma,wav) to text application would REALLY help in my job and a friendly user from Reddit pointed me to this. He tested it and told me for English at least, the results were pretty awesome!
My problem? I need Greek audio and i guess a big(?) dictionary since its a medical job, no idea if medical terms are in regular dictionaries(no latin words, since most medical words have a Greek root).
Are there any plans to add Greek? Obviously im talking about the voice recognition and text and not the software itself.
Is the software compatible with external dictionaries? For example, another software that i was looking into(does the opposite, text to audio) had a 5.5gb Greek dictionary(its a .tgz file).
Please forgive me if i asked a stupid question, but ive been searching for an offline tool like this for ages. Since its a medical job, i cant do anything online for securities reasons.

1) when I want to read the Greek NT aloud there is no option to do so. When I open a Greek NT - any of the ones I own - the options for "Read Aloud" and "Show word by word indicator" are not present. See below (I'm not sure if the screen snip is visible in the post...I can see it while editing...).

2) if I try to read anything (Say an English translation...) there is no sound at all. (The media player "box" is present...) Other media sounds (including the Logos startup sound) all work. I can play music for instance or listen/watch a Bill Mounce lecture...

For what its worth: I do not see the read aloud either. I tried to replicate KS4J's screenshot. The resource says it is downloaded in the library, but when I open the resource, I see "download all media" in the menu. When I clicked on that button and waited a while, it changed to "cancel" and "delete" buttons. When I tried to play the resource, I was told that I was offline.

I do not have super high speed internet. It appears that it was trying to download the media files, but either did not succeed or didn't finish. Since I do not use this feature, I don't think I want a gig or more of audio files... so I cancelled and deleted.

If I open the Greek Audio resource and look at the panel that opens when clicking the three dots in the top right corner of the tab - there is an option to delete the media - so I assume the media is downloaded.

Ron, are you sure that you own this product: -audio-new-testament - I think this gives a library resource of media files as well as (possibly a dataset?) the functionality of Read Aloud which shows as speaker icons on some GNTs in the library, drives the Ctrl-R thing and allows the bouncing ball. There was the case in the past that people got only the library part (i.e. that product: -greek-audio-new-testament-na27-audio ) which comes through some packages but is not the complete functionality.

My Logos library also has Westcott and Hort Greek New Testament (1881) with Morphology & The New Testament in the Original Greek: Byzantine Textform 2005 with Morphology (BYZ)(Robinson-Pierpont) that can be used with Greek Audio NT Read Aloud.

maybe there is an issue with sound that is generated / accessed with the Read Aloud Ctrl-R feature as opposed to the laying of lectures or the startup chime. I wonder if this could this be a side effect of the hardware acceleration bug? Maybe CS can help you by having a second-level technician log onto your machine.

I recommend enabling logging then start the application and make sure that if you have any pending updates, you allow those to download. If you don't have any pending updates, then wait for a few minutes (which will give the application time to complete anything it might be trying to do in the background). Then close the application and collect and post your log files here.

Your application is set to only download from the hours of midnight to 6 in the morning. Because of this setting, you aren't getting the necessary download. The application would need to be running during that range of time for the download to proceed. You can either change this setting back to Midnight to Midnight (or some other time range when you know the application will be running), or you can execute the update now command in the command box at the top of the window to just kick off the download immediately without needing to change settings.

Thanks for that. Yes read aloud works now. It seems that for Windows 11, Windows Media Player is deprecated. The default configuration for Windows Media Player in Windows 11 is for it not to be installed/enabled. Who knew?

To improve Speech to text recognition accuracy, customization is available for some languages and base models. Depending on the locale, you can upload audio + human-labeled transcripts, plain text, structured text, and pronunciation data. By default, plain text customization is supported for all available base models. To learn more about customization, see custom speech.

5 The OpenAI text to speech voices in Azure AI Speech are in public preview and only available in North Central US (northcentralus) and Sweden Central (swedencentral). Locales not listed for OpenAI voices aren't supported. For information about additional differences between OpenAI text to speech voices and Azure AI Speech text to speech voices, see OpenAI text to speech voices.

Multilingual voices can support more languages. This expansion enhances your ability to express content in various languages, to overcome language barriers and foster a more inclusive global communication environment.

2 The neural voice is a multilingual voice in Azure AI Speech. All multilingual voices can speak in the auto-detected language of the input text in the default locale without using SSML. However, you can still use the element to adjust the speaking accent of each language to set preferred accent such as British accent (en-GB) for English.

3 The OpenAI text to speech voices in Azure AI Speech are in public preview and only available in North Central US (northcentralus) and Sweden Central (swedencentral). Locales not listed for OpenAI voices aren't supported. For information about additional differences between OpenAI text to speech voices and Azure AI Speech text to speech voices, see OpenAI text to speech voices.

In some cases, you can adjust the speaking style to express different emotions like cheerfulness, empathy, and calm. All prebuilt voices with speaking styles and multi-style custom voices support style degree adjustment. You can optimize the voice for different scenarios like customer service, newscast, and voice assistant. With roles, the same voice can act as a different age and gender.

Custom neural voice lets you create synthetic voices that are rich in speaking styles. You can create a unique brand voice in multiple languages and styles by using a small set of recording data. Multi-style custom neural voices support style degree adjustment. There are two custom neural voice (CNV) project types: CNV Pro and CNV Lite (preview).

With the cross-lingual feature, you can transfer your custom neural voice model to speak a second language. For example, with the zh-CN data, you can create a voice that speaks en-AU or any of the languages with Cross-lingual support. For the cross-lingual feature, we categorize locales into two tiers: one includes source languages that support the cross-lingual feature, and the other tier comprises locales designated as target languages for cross-lingual transfer. Within the following table, distinguish locales that function as both cross-lingual sources and targets and locales eligible solely as the target locale for cross-lingual transfer.

The table in this section summarizes the 33 locales supported for pronunciation assessment, and each language is available on all Speech to text regions. Latest update extends support from English to 32 more languages and quality enhancements to existing features, including accuracy, fluency and miscue assessment. You should specify the language that you're learning or practicing improving pronunciation. The default language is set as en-US. If you know your target learning language, set the locale accordingly. For example, if you're learning British English, you should specify the language as en-GB. If you're teaching a broader language, such as Spanish, and are uncertain about which locale to select, you can run various accent models (es-ES, es-MX) to determine the one that achieves the highest score to suit your specific scenario. If you're interested in languages not listed in the following table, fill out this intake form for further assistance.

The table in this section summarizes the locales supported for Speech translation. Speech translation supports different languages for speech to speech and speech to text translation. The available target languages depend on whether the translation target is speech or text.

To set the input speech recognition language, specify the full locale with a dash (-) separator. See the speech to text language table. All languages are supported except jv-ID and wuu-CN. The default language is en-US if you don't specify a language.

To set the translation target language, with few exceptions you only specify the language code that precedes the locale dash (-) separator. For example, use es for Spanish (Spain) instead of es-ES. See the speech translation target language table below. The default language is en if you don't specify a language.

The table in this section summarizes the locales supported for Speaker recognition. Speaker recognition is mostly language agnostic. The universal model for text-independent speaker recognition combines various data sources from multiple languages. We tuned and evaluated the model on these languages and locales. For more information on speaker recognition, see the overview.

c01484d022
Reply all
Reply to author
Forward
0 new messages