--You received this message because you are subscribed to the Google Groups "Honyaku E<>J translation list" group.To unsubscribe from this group and stop receiving emails from it, send an email to honyaku+u...@googlegroups.com.To view this discussion on the web visit https://groups.google.com/d/msgid/honyaku/89425891-6568-441a-962d-82e7798a4458%40googlegroups.com.
I think "accuracy", in the context of evaluating the performance of
modern AI-based MT systems, would typically refer to a score given to
the output of such a system by a modern AI-based MT evaluation system.
Such a system (e.g. BLEU) may work, for example, on the basis of N-gram
matching to a set of reference (human) translations.
So while the idea is, on one hand, that the accuracy of machine
translation is equivalent to the closeness of its output to human
translation, an interesting feature of these metrics is that the score
for a highly skilled human translator who makes no mistakes would not
necessarily be 100% and may actually be lower than that of a good MT
system -- because the output of the highly skilled translator may be
less of an averaged match to the set of (possibly mediocre) human
translations used as a reference for calculating the score.
Thus, claims that a given MT system has achieved 99% accuracy,
human-like performance, better than human performance, etc., are not
necessarily sheer advertising puff, and may be scientifically backed --
with the (likely unmentioned) proviso that "accuracy" in this case may
have a somewhat different meaning from what would typically be
understood by that term in other contexts.
This could be called a “scientifically-backed” evaluation system, but it's a rather strange kind of science, I would say. It looks to me more like pseudo-science disguised by using a lot of fancy math to make it look like science.
The purpose of scientific activity is generally considered to be a search for truth, i.e., an understanding of reality. If it is asserted that an MT system shows a "human-like performance" and indeed may be even a "better than human performance," although we can see in experience (which is supposed to control scientific research) that MT products are often not as good as what skilled human translators can do (without considerable post-editing by humans, at least), and in many cases these translators judge that it would have been better for them to have done the translation from the start than to do the post-editing, I consider this not very congruent with reality.
I don't deny that end clients are often satisfied with MT products which are "good enough," as it is often said, for the clients' purposes. A lot of translation work that used to be done by humans when there were only human translators is now entrusted to MT, and the customers are happy with the results — they are cheap and fast. But as we know, these customers need to consider very carefully in which cases it really would be worthwhile for them to hire human translators to do the job. But these suggestions by the MT companies that their systems can replace human translators due to their tremendously superior speeds and cost reductions are the essence of false advertising, it seems to me.
The crux of the matter, I continue to believe, is that translation cannot really be done without understanding the meanings of the source texts, and computers, using AI or not, can’t possibly understand meaning. What they can do is simulate translation, and often that satisfies the customers.
Jon Johanning
Hi there,
I haven't followed this discussion closely, interesting as it is.
But now it seems to me that it is worth telling you about a
lecture at the last BDÜ conference three weeks ago (BDÜ is the
Association of Interpreters and Translators in Germany). It was
about product data sheets and a company took the trouble to train
8 different MT machines, that means to train with customer
terminology. The results were sobering for every translator. There
were different error categories from spelling to wrong translation
with appropriate weighting, and the number of errors was analysed.
In all cases, the number of errors dropped from about 20% to less
than 5% and, above all, mistranslations improved to 0%. The pie
charts previously looked green with lots of orange to red (red was
poor). Afterwards there was no red at all any more and actually
one saw essentially a green cake. Conclusion: Product data sheets
can be left to the machines with training in the future.
I agree that there are a number of types of text where MT reaches
its limits and nothing will work without qualified translators -
or rather experts in the field. Japanese will also be problematic
for a while. And especially tapeworm sentences - eternally long
sentences - will pose problems for even longer. I often find the
verb wrongly placed in posttediting. But I believe that this is
only a matter of time until enough data is available.
So Rosetta's planning 99% accuracy. Well, they will make it, at
least for certain types of text. And the verb at the end won't be
a problem.
And just for the record, this text was written in German and
translated by DEEPL. I haven't reworked anything at all that I
would do otherwise. I have already seen a few passages that I
probably would have formulated differently, for example the end of
the previous sentence. So 99% probably not, but judge the English
nevertheless times. Hm, this sentence was really creepy, but I
became more colloquial. I would be interested.
A nice day from Germany,
Christiane
--
You received this message because you are subscribed to the Google Groups "Honyaku E<>J translation list" group.
To unsubscribe from this group and stop receiving emails from it, send an email to honyaku+u...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/honyaku/CALnYxcxmFA51co3pRV7dZT%2BSvPcMeksuNiPZBFiQ4PBAWS%3DZHA%40mail.gmail.com.
Dr. Christiane
Feldmann-Leben
Chemieübersetzerdienst
Böcklerstr.
1
D-76275 Ettlingen
Phon +49 (0)7243/217218
www.chemieuebersetzerdienst.de
in...@chemieuebersetzerdienst.de
I agree that there are a number of types of text where MT reaches its limits and nothing will work without qualified translators - or rather experts in the field. Japanese will also be problematic for a while. And especially tapeworm sentences - eternally long sentences - will pose problems for even longer. I often find the verb wrongly placed in posttediting. But I believe that this is only a matter of time until enough data is available.
Yes, garbage in, garbage out is certainly something, I am seeing
quite often. And sometimes the system draws funny conclusions -
which at least gives me a good laugh sometimes. An example:
asparagus is Spargel in German, and then there is the enzyme
asparaginase which is Asparaginase in German. However, I found
"Spargelinase". I loved that. Another one was "glass sons" instead
of "glass fibers" for "fils de verre". It took me quite a few
seconds before I understood what had happened.
The point, however, is that clients can train the machines with their own material and terminology lists (and the discussions turns to implementing rules as well). And if the clients have and use good translations, done by all of us, the machine will become frighteningly good. So in the end, our own good work turns against us.
This time, all errors are mine entirely. I wrote in English. ;-)
Christiane
--
From: hon...@googlegroups.com <hon...@googlegroups.com> On Behalf Of JON JOHANNING
Forever, would be my guess. Since Google translate still does the following.
OJさんは妻らを殺したとされました。
OJ was alleged to have killed his wives.
This is an in-principle problem that can’t be fixed by “more training data”, or anything else, without doing AI right. And the giddy insanity of contemporary AI hype means that very few people even understand the phrase “doing AI right”. (FWIW, the book “Rebooting AI” (highly recommended) does talk about what it would mean to do AI right.)
David J. Littleboy
Tokyo, Japan
--
You received this message because you are subscribed to the Google Groups "Honyaku E<>J translation list" group.
To unsubscribe from this group and stop receiving emails from it, send an email to honyaku+u...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/honyaku/cb7ad5c2-eac1-e306-e56e-18a971bb348b%40lmi.net.
From: hon...@googlegroups.com <hon...@googlegroups.com> On Behalf Of JON JOHANNING
- Japanese is definitely problematic; how much longer it will be is hard to say.
Forever, would be my guess. Since Google translate still does the following.
OJさんは妻らを殺したとされました。
OJ was alleged to have killed his wives.
Yes, that’s really difficult, because if you don’t know the context you might get it wrong. In fact I asked a couple of Japanese people who didn’t know who it was about and they guessed it involved bigamy. AI is still very bad at understanding context.
It’s not that “AI is bad at understanding context” it’s that current AI explicitly refuses to even try to work on what it would mean to understand context.
The recent book “Bootstrapping AI” discusses this, and talks about what AI needs to be doing. My take is far more jaundiced than said book: the things AI is doing now can’t, in principle, do what they are claiming/trying to do, and the things that would make AI possible (common sense reasoning) are much harder than the book’s authors think.
David J. Littleboy
Tokyo, Japan
--
You received this message because you are subscribed to the Google Groups "Honyaku E<>J translation list" group.
To unsubscribe from this group and stop receiving emails from it, send an email to honyaku+u...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/honyaku/6F833150-74C3-4615-9E7E-06AB8A9F28E2%40gmail.com.