Has anyone seen the output from Rozetta and other Japanese MT vendors?

167 views
Skip to first unread message

Kevin Johnson

unread,
Dec 1, 2019, 2:41:12 PM12/1/19
to Honyaku E<>J translation list
Hi friends,

I enjoy lurking in this group and have learned a lot from your informative posts.

If you're like me, the topic of MT in general is a bit stressful. I'll give a bit of a personal prologue. Feel free to skip to "the point" below to get to the real topic. 

For me, the stressful thing about MT is more the uncertainty than anything. I've been in this industry as a full-time professional for a little less than a decade, but my workload has definitely dropped off in the last year, for the first time ever. Every other year was quite fat and satisfying -- but not this year. I've averaged about $1000 USD less per month than I'm used to.

Peppering ex-employees of my main client (a Tokyo agency) with questions has suggested that "AI" and "crowdsourcing" have both had a significant impact on business.

Anyway, I'm willing to adapt, and if MT is (or is not) the future, I'm prepared to roll with the punches, whatever that may mean. Dipping into "crowdsourcing" to beef up my client base has certainly helped, and I've clawed my way back to roughly where I used to be, although every month is a big question mark. My clients are no longer just in Japan; I'm working with companies in the UK, the US, Spain, and China. They all use different payment terms and methods. This is confusing, but better confused than poor.

So things seem to be changing, at least from my end. I'd like to assess how legitimate the MT wave is for myself. (By the way, if anyone has any advice or similar stories relevant to the above, please chime in!)

The point

This brings me to my point. Although I've seen some very interesting posts on MT generally in this forum, I'm curious whether anyone has ever seen commercial output from the big Japanese MT vendors like Rozetta. 

As you may know, Rozetta has really taken off on the stockmarket this year, and if you'd bought their stock back in early 2016 you'd have 8x what you had back then. They also made what I consider to have been the best online J-E dictionary for technical terms, 産業翻訳だよ!全員集合, although they've since ended that service (perhaps because it was enabling the competition).

Of course, the stock market is arguably a hype index more than anything.

I think we're all pretty familiar with the flaws of Google Translate's J>E output. It is easily confused by elaborate subordinate clauses before nouns, drops entire chunks of the source text if it can't squeeze them in grammatically, and retains the huge one-sentence paragraphs we all know and love in Japanese (but make for unreadable English). It has no idea how to translate mistaken or oddly used kanji or kana, and various buzzwords puzzle it as well. 

My overall impression is that modern Google Translate will often produce good, coherent English, but it will often simply ignore parts of the source text to get there. What it puts out isn't always necessarily a translation so much as a grammatical English sentence that dropped a clause or two from the Japanese.

I've also known my fair share of MT vendors, who are increasingly notorious for over-promising on quality. Maybe we're in the middle of a MT revolution or maybe a lot of it is just hype for our pair.

So which is Rozetta? Hype or the real deal? Has anybody seen the commercial stuff out of Japan? I haven't, and I'd love to know if Rozetta is really doing good work or not.

Kevin Johnson

Dan Lucas

unread,
Dec 1, 2019, 3:58:59 PM12/1/19
to Honyaku E<>J translation list
I won't comment on MT, as I have little experience in this area, although what experience I do have matches yours.

However, coming at it from a slightly different angle, don't forget that we live in a cyclical world. Economic conditions around the globe have been deteriorating. There are significant problems in China, trade friction between China and the US, and Europe has problems with Brexit, Italy and the soft German economy. The declines in Japan exports (what was it, -9.2% YoY in October 2019?) and weakness in manufacturing profits no doubt reflect these factors to some extent. We are clearly in the middle of a significant, widespread slowdown.

That means companies are starting to pull in their horns, delay what projects they think they can reasonably delay, and look for costs to cut. That implies less work for contractors and suppliers. Translation is not going to be immune to that trend in the short to medium term and has probably already been affected. The secular trend of MT may be contributing to lower demand, but if we apply Occam's razor then we should probably assign a greater impact to the influence of the downward leg of the business cycle.

Remember that Japan hasn't had a serious slowdown since, what, 2011/12? The US hasn't had a recession since 2009. We're long overdue a correction, or worse. Hold on tight.

To put it another way, you wonder in your email whether the phenomenon of the sea withdrawing from the shore might not indicate a tsunami in the offing. I suggest it's more likely to be the natural action of the tides.

Dan Lucas
--
You received this message because you are subscribed to the Google Groups "Honyaku E<>J translation list" group.
To unsubscribe from this group and stop receiving emails from it, send an email to honyaku+u...@googlegroups.com.

石崎敏明

unread,
Dec 1, 2019, 9:37:33 PM12/1/19
to hon...@googlegroups.com
Kevinさん、こんにちは。

>So which is Rozetta? Hype or the real deal?
まず、こちらの疑問に関する個人的な見解です。
個人的には、株価も展開する業務自体も過大評価とみています。
・株価:売上高、EPS、今後の業績の伸び、市場規模から判断すると、
 同業他社の翻訳センター(銘柄コード 2483)の時価総額68億と比べ
 ロゼッタの419億は正当化できない水準。
・業務:法人向けサービスは利用したことがありませんが、
   ウェブ上で展開する「アイちゃん」 https://www.traai.com/mtの質は、
   NTTドコモ傘下の「みらい翻訳」 https://miraitranslate.com/よりも低い傾向があります。

私の場合も、直接取引する資産運用会社がコスト削減の一環でMTを導入した影響のみで、
Kevinさんと同程度の減収要因が発生しました。
今、対策を検討中ですが、内容が比較的容易なマクロ経済のレポート等は、
MTから奪い返すのは不可能と割り切っています。

ファンドのパフォーマス・レポートや、
国外ファンドのピッチブック等、MTが対応できない英日翻訳はまだかなりありますので、
これから少しずつマーケティング活動にも力を入れるつもりです。

 金融翻訳  石崎敏明

Kevin Johnson

unread,
Dec 10, 2019, 8:13:43 AM12/10/19
to Honyaku E<>J translation list
Thank you both for your informative responses.

I agree that Rozetta's stock is absolutely overvalued. Regardless, I still have no idea what their product really looks like in practice. 

(There was a PR Times article from 2018 that had some side-by-side translations on it, and these were honestly terrible. But it seems Rozetta is now planning to release a "99% accurate" MT service in 2020, so we'll see whether it's snake oil or the real thing.)

Thanks for the information on other services. I have heard that the E>J market is quite different from the J>E market, so I have to wonder whether the effects on these respective markets will be different. I have heard some say that E>J machine translation is much better than J>E machine translation. I have no idea if this is true.

In any case, things are much more "feast or famine" than they used to be for me. This month is pretty good so far. We'll see, I suppose.

Thanks again,

Kevin P. Johnson

JON JOHANNING

unread,
Dec 10, 2019, 8:44:13 AM12/10/19
to Honyaku E<>J translation list
> On Dec 10, 2019, at 1:34 AM, Kevin Johnson <kpatric...@gmail.com> wrote:
>
> (There was a PR Times article from 2018 that had some side-by-side translations on it, and these were honestly terrible. But it seems Rozetta is now planning to release a "99% accurate" MT service in 2020, so we'll see whether it's snake oil or the real thing.)

I don’t know anything about Rozetta, but I do know that these “accuracy” claims made by MT companies are sheer advertising puffery. What does “accuracy” even mean in relation to these machine “translation” systems? Anyone who takes these claims seriously needs to study what MT really is and does a little more carefully.

Jon Johanning
jjoha...@igc.org

Matthew Schlecht

unread,
Dec 10, 2019, 12:35:19 PM12/10/19
to Honyaku
Indeed!
What a bean counter can mean by accuracy might be the frequency with which all the source terms are reliably represented by target terms, but the measure of the accuracy of a text should/must incorporate some "measure" of meaning. Converting Japanese into an English word salad does not represent accuracy. 
There is one (at least one) very basic problem in JA>EN MT (and a related one in DE>EN MT) with the "verb at the end" issue.
Didn't we just go through all of this about six months ago?
It is a vast oversimplification, but MT has largely been developed on the basis of English, an SVO language that identifies semantic relationships based on proximity, and the basic assumptions break down in inflected languages and SOV languages (e.g., Japanese).

MT parses segments defined by hard or soft line breaks, and occurrences of "。", and one can frequently end up with source and target segments that suffer from poor or misleading mapping.
Japanese patent authors sometimes place a parenthetical phrase in the middle of a sentence in which the text within the parentheses can contain one or more  "。".  Garden variety MT then keys on the   "。" occurrences and proceeds to break all of this up into separate segments with the result that the text strings before and after the parentheses (Which should all still constitute one sentence. One sentence that must be tied together properly.) fail to communicate with one another. In some MT systems, one can knit all these fragments back together to make a meaningful whole, but the knitting must be done by someone who understands Japanese.
Likewise, all the "wherein..." clauses in the characterizing section that are separated by hard or soft returns need to remain related to each other and to the preamble if any meaningful result is to be salvaged.
Many of these issues are also present in the application of human-operated CAT tools to JA>EN translation, but it is then up to that human (who presumably understands Japanese) to manicure the source strings so that sentences are segmented in an intact and meaningful way, and then to tease apart the English target to mimic the source Japanese phrasing and present a professional quality translation to the end client.

Matthew Schlecht, PhD
Word Alchemy Translation, Inc.
Newark, DE, USA
wordalchemytranslation.com

JON JOHANNING

unread,
Dec 10, 2019, 1:07:02 PM12/10/19
to hon...@googlegroups.com
On Dec 10, 2019, at 12:35 PM, Matthew Schlecht <matthew.f...@gmail.com> wrote:
>
> (Which should all still constitute one sentence. One sentence that must be tied together properly.)

Matthew,

Great example of putting sentences into parentheses!

Jon Johanning
jjoha...@igc.org

Herman

unread,
Dec 10, 2019, 9:50:29 PM12/10/19
to hon...@googlegroups.com
I think "accuracy", in the context of evaluating the performance of
modern AI-based MT systems, would typically refer to a score given to
the output of such a system by a modern AI-based MT evaluation system.
Such a system (e.g. BLEU) may work, for example, on the basis of N-gram
matching to a set of reference (human) translations.

So while the idea is, on one hand, that the accuracy of machine
translation is equivalent to the closeness of its output to human
translation, an interesting feature of these metrics is that the score
for a highly skilled human translator who makes no mistakes would not
necessarily be 100% and may actually be lower than that of a good MT
system -- because the output of the highly skilled translator may be
less of an averaged match to the set of (possibly mediocre) human
translations used as a reference for calculating the score.

Thus, claims that a given MT system has achieved 99% accuracy,
human-like performance, better than human performance, etc., are not
necessarily sheer advertising puff, and may be scientifically backed --
with the (likely unmentioned) proviso that "accuracy" in this case may
have a somewhat different meaning from what would typically be
understood by that term in other contexts.

Herman Kahn


Jon Johanning

unread,
Dec 10, 2019, 11:12:00 PM12/10/19
to Honyaku E<>J translation list
On Tuesday, December 10, 2019 at 9:50:29 Herman Kahn wrote:
 
I think "accuracy", in the context of evaluating the performance of
modern AI-based MT systems, would typically refer to a score given to
the output of such a system by a modern AI-based MT evaluation system.
Such a system (e.g. BLEU) may work, for example, on the basis of N-gram
matching to a set of reference (human) translations.

So while the idea is, on one hand, that the accuracy of machine
translation is equivalent to the closeness of its output to human
translation, an interesting feature of these metrics is that the score
for a highly skilled human translator who makes no mistakes would not
necessarily be 100% and may actually be lower than that of a good MT
system -- because the output of the highly skilled translator may be
less of an averaged match to the set of (possibly mediocre) human
translations used as a reference for calculating the score.

Thus, claims that a given MT system has achieved 99% accuracy,
human-like performance, better than human performance, etc., are not
necessarily sheer advertising puff, and may be scientifically backed --
with the (likely unmentioned) proviso that "accuracy" in this case may
have a somewhat different meaning from what would typically be
understood by that term in other contexts.

This could be called a “scientifically-backed” evaluation system, but it's a rather strange kind of science, I would say. It looks to me more like pseudo-science disguised by using a lot of fancy math to make it look like science.


The purpose of scientific activity is generally considered to be a search for truth, i.e., an understanding of reality. If it is asserted that an MT system shows a "human-like performance" and indeed may be even a "better than human performance," although we can see in experience (which is supposed to control scientific research) that MT products are often not as good as what skilled human translators can do (without considerable post-editing by humans, at least), and in many cases these translators judge that it would have been better for them to have done the translation from the start than to do the post-editing, I consider this not very congruent with reality.


I don't deny that end clients are often satisfied with MT products which are "good enough," as it is often said, for the clients' purposes. A lot of translation work that used to be done by humans when there were only human translators is now entrusted to MT, and the customers are happy with the results — they are cheap and fast. But as we know, these customers need to consider very carefully in which cases it really would be worthwhile for them to hire human translators to do the job. But these suggestions by the MT companies that their systems can replace human translators due to their tremendously superior speeds and cost reductions are the essence of false advertising, it seems to me.


The crux of the matter, I continue to believe, is that translation cannot really be done without understanding the meanings of the source texts, and computers, using AI or not, can’t possibly understand meaning. What they can do is simulate translation, and often that satisfies the customers.


Jon Johanning

jjoha...@igc.org

Christiane Feldmann-Leben

unread,
Dec 11, 2019, 3:27:02 AM12/11/19
to hon...@googlegroups.com

Hi there,

I haven't followed this discussion closely, interesting as it is. But now it seems to me that it is worth telling you about a lecture at the last BDÜ conference three weeks ago (BDÜ is the Association of Interpreters and Translators in Germany). It was about product data sheets and a company took the trouble to train 8 different MT machines, that means to train with customer terminology. The results were sobering for every translator. There were different error categories from spelling to wrong translation with appropriate weighting, and the number of errors was analysed. In all cases, the number of errors dropped from about 20% to less than 5% and, above all, mistranslations improved to 0%. The pie charts previously looked green with lots of orange to red (red was poor). Afterwards there was no red at all any more and actually one saw essentially a green cake. Conclusion: Product data sheets can be left to the machines with training in the future.
I agree that there are a number of types of text where MT reaches its limits and nothing will work without qualified translators - or rather experts in the field. Japanese will also be problematic for a while. And especially tapeworm sentences - eternally long sentences - will pose problems for even longer. I often find the verb wrongly placed in posttediting. But I believe that this is only a matter of time until enough data is available.
So Rosetta's planning 99% accuracy. Well, they will make it, at least for certain types of text. And the verb at the end won't be a problem.
And just for the record, this text was written in German and translated by DEEPL. I haven't reworked anything at all that I would do otherwise. I have already seen a few passages that I probably would have formulated differently, for example the end of the previous sentence. So 99% probably not, but judge the English nevertheless times. Hm, this sentence was really creepy, but I became more colloquial. I would be interested.

A nice day from Germany,
Christiane



--
You received this message because you are subscribed to the Google Groups "Honyaku E<>J translation list" group.
To unsubscribe from this group and stop receiving emails from it, send an email to honyaku+u...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/honyaku/CALnYxcxmFA51co3pRV7dZT%2BSvPcMeksuNiPZBFiQ4PBAWS%3DZHA%40mail.gmail.com.
--



Dr. Christiane Feldmann-Leben
Chemieübersetzerdienst
Böcklerstr. 1
D-76275 Ettlingen

Phon +49 (0)7243/217218
www.chemieuebersetzerdienst.de
in...@chemieuebersetzerdienst.de



Mark Spahn

unread,
Dec 11, 2019, 6:23:20 AM12/11/19
to hon...@googlegroups.com
- - - - - -

Hi Christiane.  It would have been interesting if you had included your
original German-text note, so that your readers can compare it with its
MT version.

-- Mark Spahn (West Seneca, NY)


Herman

unread,
Dec 11, 2019, 6:30:48 AM12/11/19
to hon...@googlegroups.com
On 12/11/19 12:26 AM, Christiane Feldmann-Leben wrote:

> Well, they will make it, at least for certain types of text. And the
> verb at the end won't be a problem. And just for the record, this
> text was written in German and translated by DEEPL. I haven't
> reworked anything at all that I would do otherwise. I have already
> seen a few passages that I probably would have formulated
> differently, for example the end of the previous sentence. So 99%
> probably not, but judge the English nevertheless times. Hm, this
> sentence was really creepy, but I became more colloquial. I would be
> interested.
>

Well, I was interested, so as an experiment, I tried running the above
text through DEEPL, translating it back and forth between different
languages that I have some knowledge of, but then it said that "Due to
high traffic on the free DeepL translator, the service is limited at the
moment.", so I switched to Google Translate. Anyhow, things were going
relatively well in the sense that the meaning of the text was being more
or less maintained, albeit with some degradation. But then I threw Latin
into the mix (which I did "take" at one point, although my knowledge of
it now is, admittedly, rather poor), and that sort of messed some of the
text up into semi-gibberish, leading back to the following English:

"Well, at least some organization types can be supported. The last word
is that it doesn't matter. For information translated by DEEPL to be in
German. Otherwise I don't have to start over. The reason why you can't
see what you can't do is, as an example, another way, at the end of the
grace of the previous judgment is already past. 99% due to UK party
oblivion. This sentence is scary, but at most. I am interested"


However, further iterations seemed to improve the quality of the text
into something that made sense again:

"Well, at least some types of organizations are compatible. The last
suggestion is that it doesn't matter. Use DEEPL for German translation.
Otherwise, there is no need to start over from the beginning. For
example, he doesn't see what he can't do because it happened at the end
of a pre-estimated grace period. 99% due to forgetting things in the UK.
This prayer is terrible. be interested"

Now, on the one hand, this sort of underhanded activity could be, for
example, leveraged to impeach the very process of machine translation as
being inherently unreliable, but on the other hand, the process of going
through a number of machine translation iterations between a handful of
languages seemed to separate the wheat from the chaff, bringing into
clear relief what in retrospect may have been the underlying intended
meaning or message of the initial text: "Use DEEPL for German translation."

Herman Kahn


JON JOHANNING

unread,
Dec 11, 2019, 10:22:14 AM12/11/19
to hon...@googlegroups.com
On Dec 11, 2019, at 3:26 AM, Christiane Feldmann-Leben <cle...@web.de> wrote:

I agree that there are a number of types of text where MT reaches its limits and nothing will work without qualified translators - or rather experts in the field. Japanese will also be problematic for a while. And especially tapeworm sentences - eternally long sentences - will pose problems for even longer. I often find the verb wrongly placed in posttediting. But I believe that this is only a matter of time until enough data is available.

Product data sheets are a good example of texts that used to be available to human translators when there were only human translators, but much less so now. Japanese is definitely problematic; how much longer it will be is hard to say. And “tapeworm sentences” (love that term, Christiane!) are exactly what texts such as patent applications are full of.

“Enough data” is a real problem for MT development. I’m not sure that the developers have yet solved the famous problem we have heard of in the past: the enormous amounts of translation data needed to train the systems is usually scraped from web sites, which are often translated by raw MT, without expert post-editing, so it’s a garbage-in, garbage-out situation.

Jon Johanning

Christiane Feldmann-Leben

unread,
Dec 11, 2019, 10:43:42 AM12/11/19
to hon...@googlegroups.com

Yes, garbage in, garbage out is certainly something, I am seeing quite often. And sometimes the system draws funny conclusions - which at least gives me a good laugh sometimes. An example: asparagus is Spargel in German, and then there is the enzyme asparaginase which is Asparaginase in German. However, I found "Spargelinase". I loved that. Another one was "glass sons" instead of "glass fibers" for "fils de verre". It took me quite a few seconds before I understood what had happened.

The point, however, is that clients can train the machines with their own material and terminology lists (and the discussions turns to implementing rules as well). And if the clients have and use good translations, done by all of us, the machine will become frighteningly good. So in the end, our own good work turns against us.

This time, all errors are mine entirely. I wrote in English. ;-)

Christiane

--

dav...@gol.com

unread,
Dec 11, 2019, 2:16:06 PM12/11/19
to hon...@googlegroups.com

 

From: hon...@googlegroups.com <hon...@googlegroups.com> On Behalf Of JON JOHANNING

  • Japanese is definitely problematic; how much longer it will be is hard to say.

 

Forever, would be my guess. Since Google translate still does the following.

 

OJさんは妻らを殺したとされました。

OJ was alleged to have killed his wives.

 

This is an in-principle problem that can’t be fixed by “more training data”, or anything else, without doing AI right. And the giddy insanity of contemporary AI hype means that very few people even understand the phrase “doing AI right”. (FWIW, the book “Rebooting AI” (highly recommended) does talk about what it would mean to do AI right.)

 

David J. Littleboy

Tokyo, Japan

 

 

Kevin Johnson

unread,
Dec 11, 2019, 3:15:19 PM12/11/19
to Honyaku E<>J translation list
Great post, Christiane! It was like a short story with the reveal at the end...

There are a couple things that are nagging me here. One, my understanding is that DeepL is trained on the EU corpus and therefore has absolute reams of high-quality side-by-side translations to work with. Meanwhile, as we all know, the full of body of Japanese-English work is garbage, and your average Japanese academic paper (or company website, or even your average Japanese-English dictionary, etc.) has a terrible English translation. So the texts may not be there (but, conversely, standards may be quite low).

My second concern is that, as far as I know, German and English are extremely similar languages, making the task of automating such translations much easier. My understanding is that it's been much easier for Europe to make use of MT, and a ton of the huge success stories I hear are in the German-English pair specifically. I'm sure there the EU corpus enables similar results for many European pairs, especially closely related ones.

Anyway, I'd love to know if anybody knows of similar success stories specific to Japanese-English (which is altogether different from English-Japanese). I have never heard one, but I'm eager to know. I don't view MT as the enemy and welcome disruptive change - perhaps I'm a masochist :)

Kevin Johnson

Herman

unread,
Dec 11, 2019, 5:52:26 PM12/11/19
to hon...@googlegroups.com
On 12/11/19 8:09 AM, Kevin Johnson wrote:
> Great post, Christiane! It was like a short story with the reveal at the
> end...
>
> There are a couple things that are nagging me here. One, my
> understanding is that DeepL is trained on the EU corpus and therefore
> has absolute reams of high-quality side-by-side translations to work
> with. Meanwhile, as we all know, the full of body of Japanese-English
> work is garbage, and your average Japanese academic paper (or company
> website, or even your average Japanese-English dictionary, etc.) has a
> terrible English translation. So the texts may not be there (but,
> conversely, standards may be quite low).
>
> My second concern is that, as far as I know, German and English are
> extremely similar languages, making the task of automating such
> translations much easier. My understanding is that it's been much easier
> for Europe to make use of MT, and a ton of the huge success stories I
> hear are in the German-English pair specifically. I'm sure there the EU
> corpus enables similar results for many European pairs, especially
> closely related ones.
>

There is no single criterion of similarity between languages to begin
with, and plus the question of what level of similarity is to be
characterized as extreme vs slight is somewhat subjective and depends
also on what other pairs of languages one is comparing the given pair
to. I would say, though, that German and English are not so similar that
the translation between German and English would necessarily be easier
than, say, between English and Japanese, simply by virtue of greater
similarity at, e.g., morphological, syntactic and lexical levels. I
would say that the greater success that has apparently been had with MT
of languages based on the EU corpus (and the like) may derive more from
similarity of what is said than similarity in the means of saying it. In
other words, it is not the case that speakers of different languages
necessarily say 100% the same things but only do so in a differently
encoded manner, but rather, for various reasons -- including not only
differences of a phonological, morphological, syntactic or other such
purely linguistic nature, but also differences of a historical,
cultural, political, economic, philosophical, etc., nature -- speakers
of different languages may say different kinds of things.

Even if the languages are extremely different at a purely linguistic
encoding level, provided that a sufficient pattern of correspondences
can be established on the basis of an adequate bilingual corpus to begin
with, this type of difference would not be a source of difficulty for an
MT system. The difficulty would come from an absence of 100% or nearly
100% corresponding terms, expressions, etc., not from the fact that the
equivalent terms and expressions are formally dissimilar. This point
would be exemplified, for instance, by comparing the case of
(1) translating the language of a Chinese-speaking practitioner of
Western medicine into language of an English-speaker practitioner of
Western medicine, versus (2) the case of translating the language of a
Chinese-speaking practitioner of Chinese medicine into the language of
an English-speaking practitioner of Western medicine. Assuming that the
source language content remains invariant, the only way to establish a
system of translation in case (2) that would work as well as in case (1)
would be to in effect devise a consistent body of "terrible English
translation", and then accept that pattern of terrible English
translation as the correct translation.


Herman Kahn

石崎敏明

unread,
Dec 12, 2019, 7:07:58 AM12/12/19
to hon...@googlegroups.com
Jon Johanning writes:

<snip>
“Enough data” is a real problem for MT development. I’m not sure that the developers have yet solved the famous problem we have heard of in the past: the enormous amounts of translation data needed to train the systems is usually scraped from web sites, which are often translated by raw MT, without expert post-editing, so it’s a garbage-in, garbage-out situation.

私はAI開発や著作権の詳しい知識がありませんが、無知を覚悟で一言、二言。
仮に、データ・ソースがウェブサイトでなく書籍だったら、garbage-in, garbage-outにはなりませんよね。
例えば、GoogleがBig Shortのオリジナルと日本語訳の両方をインプットして、MTに活用したとすると、
ゴミどころか、Beyond Meatのような商品性のある製品を複製することになります。

こうなると、「金融・資産運用の オタク的な知識」という囲いの中にある
「人間保護区域」に住んでいる私は、柵を失ってロボットに対し全く無防備になります。

結論:書籍データの扱い次第では、今、そこにある危機はロゼッタではなくGoogleである可能性も。

石崎敏明

2019年12月12日(木) 7:52 Herman <sl...@lmi.net>:
--
You received this message because you are subscribed to the Google Groups "Honyaku E<>J translation list" group.
To unsubscribe from this group and stop receiving emails from it, send an email to honyaku+u...@googlegroups.com.

Herman

unread,
Dec 12, 2019, 12:00:20 PM12/12/19
to hon...@googlegroups.com
On 12/12/19 4:07 AM, 石崎敏明 wrote:
> Jon Johanning writes:
>
> <snip>
> “Enough data” is a real problem for MT development. I’m not sure that
> the developers have yet solved the famous problem we have heard of in
> the past: the enormous amounts of translation data needed to train the
> systems is usually scraped from web sites, which are often translated by
> raw MT, without expert post-editing, so it’s a garbage-in, garbage-out
> situation.
>
> 私はAI開発や著作権の詳しい知識がありませんが、無知を覚悟で一言、二言。
> 仮に、データ・ソースがウェブサイトでなく書籍だったら、garbage-in,
> garbage-outにはなりませんよね。
> 例えば、GoogleがBig Shortのオリジナルと日本語訳の両方をインプットして、
> MTに活用したとすると、
> ゴミどころか、Beyond Meatのような商品性のある製品を複製することになります。
>
> こうなると、「金融・資産運用の オタク的な知識」という囲いの中にある
> 「人間保護区域」に住んでいる私は、柵を失ってロボットに対し全く無防備にな
> ります。
>
> 結論:書籍データの扱い次第では、今、そこにある危機はロゼッタではなく
> Googleである可能性も。
>


Garbage-in, garbage-out
といっても、ソースデータが翻訳ミスをたくさん含んでいるのがGarbage、とは限らない。

例えば、

「年末やお正月のすき焼きにたっぷりの笹がきごぼうをぜひ加えてさっと煮て下さい。」

といったレシピからの一文に対して、Google Translateが

"Be sure to add plenty of mushrooms to the end of the year and New
Year's sukiyaki and simmer."

との結果を出す根拠は、「笹がきごぼう」を
mushroomsにした誤訳がインターネット上に多い、ということではなく、「すきやきにたっぷりのきのこを」「きのこたっぶりのすきやき」などの文がインターネット上に多いからである。

即ち、「すきやきにたっぷりのきのこ」等が、「すき焼きにたっぷりの笹がきごぼう」に対してGarbageとなる。

Herman Kahn

石崎敏明

unread,
Dec 12, 2019, 3:56:05 PM12/12/19
to hon...@googlegroups.com
Herman Kahn writes:

との結果を出す根拠は、「笹がきごぼう」を
mushroomsにした誤訳がインターネット上に多い、という
ことではなく、「すきやきにたっぷりのきのこを」「きのこたっぶりのすきやき」などの文がインターネット上に多いからである。

即ち、「すきやきにたっぷりのきのこ」等が、「すき焼きにたっぷ
りの笹がきごぼう」に対してGarbageとなる。

=====
歯車がかみあっていませんね。

結論で書いたように、 私は書籍のデータに限定しているんですけど (^^;

石崎敏明

Herman

unread,
Dec 12, 2019, 4:56:02 PM12/12/19
to hon...@googlegroups.com
だから、訓練データとなる書籍において、「すきやき たっぷり きのこ」が関連付けられている文が多く、「すきやき たっぷり 笹掻きごぼう」が一緒に出る文が無いかごく少ない場合、Google
Translate
のようなAIを使ったMTならば、笹掻きごぼう≒きのこ、という関連付けができて、笹掻きごぼうをmushroomsと訳してしまう可能性が高い。即ち、対象文が訓練データのパターンに合わない場合です。

Herman Kahn


Jens Wilkinson

unread,
Dec 12, 2019, 6:11:20 PM12/12/19
to hon...@googlegroups.com

On Dec 12, 2019, at 4:16, dav...@gol.com wrote:



 

From: hon...@googlegroups.com <hon...@googlegroups.com> On Behalf Of JON JOHANNING

  • Japanese is definitely problematic; how much longer it will be is hard to say.

 

Forever, would be my guess. Since Google translate still does the following.

 

OJさんは妻らを殺したとされました。

OJ was alleged to have killed his wives.

 


Yes, that’s really difficult, because if you don’t know the context you might get it wrong. In fact I asked a couple of Japanese people who didn’t know who it was about and they guessed it involved bigamy. AI is still very bad at understanding context. 

Jens Wilkinson

Herman

unread,
Dec 12, 2019, 7:42:59 PM12/12/19
to hon...@googlegroups.com
On 12/12/19 3:11 PM, Jens Wilkinson wrote:
>
>> On Dec 12, 2019, at 4:16, dav...@gol.com wrote:
>>
>> 
>>
>> *From:* hon...@googlegroups.com <hon...@googlegroups.com> *On Behalf
>> Of *JON JOHANNING
>>
>> * Japanese is definitely problematic; how much longer it will be is
>> hard to say.
>>
>> Forever, would be my guess. Since Google translate still does the
>> following.
>>
>> OJさんは妻らを殺したとされました。
>>
>> OJ was alleged to have killed his wives.
>>
>>
> Yes, that’s really difficult, because if you don’t know the context you
> might get it wrong. In fact I asked a couple of Japanese people who
> didn’t know who it was about and they guessed it involved bigamy. AI is
> still very bad at understanding context.
>

Well, if a couple of Japanese people understood it that way, that would
suggest that the AI got it right in this case. My sense is that nearly
all instances of "妻らを" that may be found in the corpus would be intended
to signify "wife and others", but in the specific case above, the
wording 「OJさんは...とされました」would seem to imply a certain specificity
that would be incompatible with reading 妻ら as "his wife and other
unspecified persons", thus leading to the interpretation "wives".

I would not say that AI is bad at understanding context, though.
Strictly speaking, AI doesn't really understand anything, but if one
takes understanding context to mean translating the same term
differently depending on the context, then that is precisely what AI
based MT systems tend to do. As an example, from Google Translate:

OJさんは妻らない = OJ doesn't have a wife
PKさんは妻らない = PK is not wife


Herman Kahn

dav...@gol.com

unread,
Dec 12, 2019, 8:08:38 PM12/12/19
to hon...@googlegroups.com

 

It’s not that “AI is bad at understanding context” it’s that current AI explicitly refuses to even try to work on what it would mean to understand context.

 

The recent book “Bootstrapping AI” discusses this, and talks about what AI needs to be  doing. My take is far more jaundiced than said book: the things AI is doing now can’t, in principle, do what they are claiming/trying to do, and the things that would make AI possible (common sense reasoning) are much harder than the book’s authors think.

 

David J. Littleboy

Tokyo, Japan

 

 

--

You received this message because you are subscribed to the Google Groups "Honyaku E<>J translation list" group.
To unsubscribe from this group and stop receiving emails from it, send an email to honyaku+u...@googlegroups.com.

Reply all
Reply to author
Forward
0 new messages