{"id":122,"date":"2006-07-18T03:55:00","date_gmt":"2006-07-17T18:55:00","guid":{"rendered":"https:\/\/fugutabetai.com\/blog\/2006\/07\/18\/notes-from-monday-2007-07-17s-talks-at-acl-coliong-2006\/"},"modified":"2006-07-18T03:55:00","modified_gmt":"2006-07-17T18:55:00","slug":"notes-from-monday-2007-07-17s-talks-at-acl-coliong-2006","status":"publish","type":"post","link":"https:\/\/fugutabetai.com\/blog\/2006\/07\/18\/notes-from-monday-2007-07-17s-talks-at-acl-coliong-2006\/","title":{"rendered":"Notes from Monday 2007-07-17&#8217;s talks at ACL\/COLIONG 2006"},"content":{"rendered":"<p>Notes from the second machine translation session at COLING\/ACL in Sydney Australia.  If you aren&#8217;t a computational linguist, this will probably not interest you.  Even if you are, I am not making any promises&#8230;<\/p>\n<p><!-- readmore --><\/p>\n<p>Machine Translation Session I:<\/p>\n<p>Combining Arabic Preprocessing Schemes for Statistical Machine Translation<br \/>\nFatiha Sadat, Nizar Habash<\/p>\n<p>The presentation was not very polished, and slides had lots of text on<br \/>\nthem.  The content seemed very nice though.  <\/p>\n<p>Nizar is not here, but I did run into Owen Rambow later.  <\/p>\n<p>&#8212;<\/p>\n<p>Going Beyond AER: An Extensive Analysis of Word Alignment and Their Impact on MT<br \/>\nNecip Fazil Ayan, Bonnie J. Dohr<\/p>\n<p>I thought this was a good presentation, but Bonnie went very quickly<br \/>\non some of the slides.  Overall if you are using BLEU or Alignment<br \/>\nError Rate (AER) for evaluation, you have to look at the actual<br \/>\napplication to be able to make a good evaluation.  AER doesn&#8217;t seem to<br \/>\nbe a very good metric for word alignment evaluation for phrase<br \/>\ntranslation.  <\/p>\n<p>&#8212;<\/p>\n<p>Discriminative Word Alignment with Conditional Random Fields<br \/>\nPhil Blunsom, Trevor Cohn (The University of Melbourne)<\/p>\n<p>Because they do forward and backwards models, they can do many to many<br \/>\n(since one direction can do one to many) alignments.  <\/p>\n<p>They add features to the alignment process that you can&#8217;t use in<br \/>\nGiza.  Orthographic features like (stems, prefix and suffix match,<br \/>\nword length difference.)  Markov feature, lets model learn that one to<br \/>\nmany are not good?  Relative sentence position, POS tags, bilingual<br \/>\ndictionary, some features on null.  <\/p>\n<p>They conclude that Conditional Random Fields are a very good model for<br \/>\nword alignment even with small number of aligned words per sentence.<br \/>\nThey are tractable to train and use.  Using multiple features are<br \/>\nuseful, especially when combined with using other models (GIZA \/ IBM<br \/>\nmodel 4) are very useful.  Markov sequence features improve<br \/>\nalignments.  <\/p>\n<p>&#8212;<\/p>\n<p>Named Entity Transliteration with Comparable Corpora<br \/>\nRichard Sproat, Tao Tao, ChengXiang Zhai (University of Illinois at<br \/>\nUrbana-Champaign)<\/p>\n<p>Using data from Xinhau in English and Chinese from the same news<br \/>\nagency.  Start with an English name, and try to identify Chinese<br \/>\ncharacter n-grams.  Score the candidates via phonetic scoring, and a<br \/>\nfrequency profile looking at how the English and Chinese candidate<br \/>\nshow up over time together.  Looks also at scores based on<br \/>\ncooccurrences in document pairs over corpora.  <\/p>\n<p>Phonetic model is probabilistic noisy source-channel model that maps<br \/>\nfrom Chinese characters into sequences of English phonemes.  Uses<br \/>\nGood-Turning smoothing for unseen mappings.  <\/p>\n<p>Identify texts with potential names by searching for strings with<br \/>\nsequences of characters (about 500) that are used for transliteration<br \/>\ncommonly.  <\/p>\n<p>If you have documents that contain other potential transliteration<br \/>\npairs that are in the same document, you increase your weight for the<br \/>\ncandidate.  They introduce a new parameter to account for scores being<br \/>\nreinforced by co-occurring terms also being likely translation pairs.<br \/>\nSeems to be similar to PageRank (they say, and it makes sense.) <\/p>\n<p>Both the phonetic and time based model are known I think, but the<br \/>\nscore propagation is novel, and it is shown to help.  <\/p>\n<p>At the talk, Pascale Fung was sitting next to me, but I didn&#8217;t notice<br \/>\nuntil mid-way through the talk.  She suggests using the time frequency<br \/>\nmeasure as a pre-selection step for generating candidates.  <\/p>\n<p>Also asked a question on whether they have compared their approach to<br \/>\nusing a large dictionary.  <\/p>\n<p>&#8212;<\/p>\n<p>Extracting Parallel Fragments from Comparable Corpora<br \/>\nDragos Munteanu, Daniel Marcu<\/p>\n<p>Parallel data is almost entirely from the Political domain<br \/>\n(parliamentary data) and news.  Looks like they are doing comparable<br \/>\ncorpora extraction from the web.  They are focusing on finding<br \/>\nparallel bits from the document pairs, but are not talking about<br \/>\nfinding candidate documents.  <\/p>\n<p>They use a bilingual lexicon in their system that has three features:<br \/>\nhigh precision, probability of translation, probability of not being<br \/>\ntranslations.  They learn this dictionary, using log likelyhood ratio<br \/>\nscores, and put a positive or negative association on them based on<br \/>\noccurrence patterns.  <\/p>\n<p>Using this learned resource they find sub-fragments of a sentence that<br \/>\nare translations of each other.  They look at a smoothed signal<br \/>\n(average translation probabilities of word + five surrounding areas)<br \/>\nto see which parts of the sentence are translation.  <\/p>\n<p>They evaluate comparing MT systems that are training on initial data,<br \/>\nor data + extracted parallel fragments, data + extracted sentences.  <\/p>\n<p>(In general, the font was so small on the slides that I could not read<br \/>\nany of the examples at all.)<\/p>\n<p>Question from Pascale about using Mutual Information with T-score to<br \/>\nsolve the problem with fewer steps instead of log likelyhood ratio.  <\/p>\n<p>&#8212;<\/p>\n<p>Applications I session<\/p>\n<p>Automated Japanese Essay Scoring System based on Articles written by<br \/>\nExperts<br \/>\nTsunenori Ishioka, Masayuki Kameda<\/p>\n<p>Their system rates on rhetoric, organization, and contents.  <\/p>\n<p>Rhetorical: They have some metrics for rhetorical judgments: shorter<br \/>\nsentence length, clause length, number of clauses, kanji \/ kana ratio,<br \/>\nnumber of embeddings, vocabulary diversity, converts kanji to kana<br \/>\nreading then counts lengths where longer is more complicated, and<br \/>\npassive sentences.  They compute statistical distributions from<br \/>\neditorials and text from Mainichi news.  They compare input text to<br \/>\nthe distribution over these variables.<\/p>\n<p>Organization: They have a list of (125) connective conjunctive phrases<br \/>\nthat are used to segment the document, they are classified into 8<br \/>\ncategories.  They look at the number of connective conjunctions, and<br \/>\nlook at the order of the appearance with a trigram and unigram model.  <\/p>\n<p>Latent Semantic Indexing is used to evaluate contents.  Their system<br \/>\nworks on essays from 400-800 characters in Japanese.  <\/p>\n<p>&#8212;<\/p>\n<p>A Feedback-Augmented Method for Detecting Errors in the Writing of<br \/>\nLearners of English<br \/>\nRyo Nagata, Atsuo Kawaii, Koichiro Morihiro, Naoki Isu<br \/>\n(Hyogo University of Teacher Education and Mie University)<\/p>\n<p>They are focusing on errors of articles and countability. <\/p>\n<p>They generate WIC lists for a word with context to get examples for<br \/>\nrules, then have some rule based system for classifying into countable<br \/>\nand uncountable instances.  They then learn rules to predict<br \/>\ncountability, and learn a log liklihood ratio to determine when to<br \/>\napply the rules.  The rules are predicated on the context of the word.  <\/p>\n<p>Augmenting the decision lists with feedback.  They use a corpus of<br \/>\nteachers&#8217; corrections to learn rules, and a probabilistic model to mix<br \/>\nthem with the rules from the general corpus.  Since the feedback<br \/>\ncorpus is small, they looked at many types of models for mixing the<br \/>\nresults.  <\/p>\n<p>I think it would have been nice to see an example of the feedback in<br \/>\nthe feedback corpus.  <\/p>\n<p>Question about the statistical significance of the results &#8211; they were<br \/>\nnot statistically significant across different types of the system.<br \/>\nAnother question about corpus choice, perhaps should download texts on<br \/>\nthe specific topic to improve lexical coverage.  Comment that he<br \/>\nshould look at Bond&#8217;s work on AltJ\/E and some of his papers on<br \/>\ntranslation and countability. <\/p>\n<p>&#8212;<\/p>\n<p>Correcting ESL Errors using Phrasal SMT Techniques<br \/>\nChris Brokett, William B. Dolan, Michael Gamon<br \/>\n(Microsoft Research)<\/p>\n<p>The idea of ESL error correction as statistical machine translation.<br \/>\nThey model the learner errors as a noisy channel model and try to find<br \/>\nthe best translation sentence with the input error sentence.  This<br \/>\npaper is a small pilot study with an off the shelf SMT system<br \/>\ncorrecting errors in mass nouns.  Used &#8220;Tree to string&#8221; system<br \/>\n(Menezes and Quirk 2005) uses unlabeled dependency parse.  <\/p>\n<p>Created artificial error-full sentences.  With 45K words of training<br \/>\ndata, they were able to correct completely about 55% of the testing<br \/>\ndata.  <\/p>\n<p>Overall I really liked this presentation.  <\/p>\n<p>&#8212;<\/p>\n<p>Dialogue II Session<\/p>\n<p>Learning to Generate Naturalistic Utterances Using Reviews in Spoken<br \/>\nDialogue Systems<br \/>\nRyuichiro Higashinaka, Rashimi Prasad, Marilyn Walker<\/p>\n<p>I enjoyed this paper.  The basic approach was to mine review websites<br \/>\n(food in this case) for reviews and their associated scores (1-5) on a<br \/>\nvariety of features (food quality, ambiance, value, etc.)  They<br \/>\ndetermined which sentences are related to the scores using a<br \/>\nhand-crafted lexicon of words that connect to the various scored<br \/>\nfeatures, and built a template that realized the scores.  One template<br \/>\ncan have multiple value associated with it, and will only be used for<br \/>\ngeneration when all the values are available to generate.  He<br \/>\ndemonstrated that the system could be plugged into a dialogue system.<br \/>\nThere were some questions that were not quite on point I think (can<br \/>\nyou use multiple websites? &#8212; of course, if you write a front-end<br \/>\nparser to scrape it, and have a good lexicon for the feature mapping.<br \/>\nAnother was an interesting question, but not on point either really.<br \/>\nThe question was a comment that the learned templates have words<br \/>\n(adjectives, descriptions) that are used for different scores.  For<br \/>\nexample, 1 and 2 for food both had &#8220;bad food.&#8221;  Of course, that is a<br \/>\nquestion of human interpretation, and since humans used these words to<br \/>\ndescribe the food, of course it is valid.)<\/p>\n<p>&#8212;<\/p>\n<p>Poster Sessions<\/p>\n<p>I saw a good poster from Satoshi Sekine from NYU.  <\/p>\n","protected":false},"excerpt":{"rendered":"<p>Notes from the second machine translation session at COLING\/ACL in Sydney Australia. If you aren&#8217;t a computational linguist, this will probably not interest you. Even if you are, I am not making any promises&#8230; Machine Translation Session I: Combining Arabic Preprocessing Schemes for Statistical Machine Translation Fatiha Sadat, Nizar Habash The presentation was not very [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[10],"tags":[],"_links":{"self":[{"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/posts\/122"}],"collection":[{"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/comments?post=122"}],"version-history":[{"count":0,"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/posts\/122\/revisions"}],"wp:attachment":[{"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/media?parent=122"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/categories?post=122"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/fugutabetai.com\/blog\/wp-json\/wp\/v2\/tags?post=122"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}