BEGIN:VCALENDAR VERSION:2.0 PRODID:-//128.220.36.25//NONSGML kigkonsult.se iCalcreator 2.26.9// CALSCALE:GREGORIAN METHOD:PUBLISH X-FROM-URL:https://www.clsp.jhu.edu X-WR-TIMEZONE:America/New_York BEGIN:VTIMEZONE TZID:America/New_York X-LIC-LOCATION:America/New_York BEGIN:STANDARD DTSTART:20231105T020000 TZOFFSETFROM:-0400 TZOFFSETTO:-0500 RDATE:20241103T020000 TZNAME:EST END:STANDARD BEGIN:DAYLIGHT DTSTART:20240310T020000 TZOFFSETFROM:-0500 TZOFFSETTO:-0400 RDATE:20250309T020000 TZNAME:EDT END:DAYLIGHT END:VTIMEZONE BEGIN:VEVENT UID:ai1ec-21615@www.clsp.jhu.edu DTSTAMP:20240329T093737Z CATEGORIES;LANGUAGE=en-US:Student Seminars CONTACT: DESCRIPTION:Abstract\n\n\nWe consider a problem of data collection for sema ntically rich NLU tasks\, where detailed semantics of documents (or uttera nces) are captured using a complex meaning representation. Previously\, d ata collection for such tasks was either handled at the cost of extensive annotator training (e.g. in FrameNet or PropBank) or simplified meaning re presentation (e.g. in QA-SRL or Overnight). In this talk\, we present two systems [1\, 2] that aim to support fast\, accurate\, and expressive sema ntic annotations by pairing human workers with a trained model in the loop .\n\nThe first system\, called Guided K-best [1]\, is an annotation toolki t for conversational semantic parsing. Instead of typing annotations from scratch\, data specialists choose a correct parse from the K-best output of a few-shot prototyped model. As the K-best list can be large (e.g. K=1 00)\, we guide the annotators’ exploration of the K-best list via explaina ble hierarchical clustering. In addition\, we experiment with RoBERTa-bas ed reranking of the K-best list to recalibrate the few-shot model towards Accuracy@K. The final system allows to annotate data up to 35% faster tha n the standard\, non-guided K-best and improves the few-shot model’s top-1 accuracy by up to 18%. The second system\, called SchemaBlocks [2]\, is an annotation toolkit for schemas\, or structured descriptions of frequent real-world scenarios (e.g.\, cooking a meal). It represents schemas in t he annotation UI as nested blocks. Using a novel Causal ARM model\, we fu rther speed up the annotation process and guide data specialists towards e xpressive and diverse schemas. As part of this work\, we collect 232 sche mas\, evaluating their internal coherence and their coverage on large-scal e newswire corpora.\n\n\n DTSTART;TZID=America/New_York:20220311T120000 DTEND;TZID=America/New_York:20220311T131500 LOCATION:Virtual Seminar SEQUENCE:0 SUMMARY:Student Seminar – Anton Belyy “Systems for Human-AI Cooperation on Collecting Semantic Annotations” URL:https://www.clsp.jhu.edu/events/student-seminar-anton-belyy-systems-for -human-ai-cooperation-on-collecting-semantic-annotations/ X-COST-TYPE:free X-ALT-DESC;FMTTYPE=text/html:\\n\\n
\\nAbstr act
\n\n X-TAGS;LANGUAGE=en-US:2022\,Belyy\,March END:VEVENT BEGIN:VEVENT UID:ai1ec-22395@www.clsp.jhu.edu DTSTAMP:20240329T093737Z CATEGORIES;LANGUAGE=en-US:Seminars CONTACT: DESCRIPTION:Abstract\nRecursive calls over recursive data are widely useful for generating probability distributions\, and probabilistic programming allows computations over these distributions to be expressed in a modular and intuitive way. Exact inference is also useful\, but unfortunately\, ex isting probabilistic programming languages do not perform exact inference on recursive calls over recursive data\, forcing programmers to code many applications manually. We introduce a probabilistic language in which a wi de variety of recursion can be expressed naturally\, and inference carried out exactly. For instance\, probabilistic pushdown automata and their gen eralizations are easy to express\, and polynomial-time parsing algorithms for them are derived automatically. We eliminate recursive data types usin g program transformations related to defunctionalization and refunctionali zation. These transformations are assured correct by a linear type system\ , and a successful choice of transformations\, if there is one\, is guaran teed to be found by a greedy algorithm. I will also describe the implement ation of this language in two phases: first\, compilation to a factor grap h grammar\, and second\, computing the sum-product of the factor graph gra mmar.\n\nBiography\nDavid Chiang (PhD\, University of Pennsylvania\, 2004) is an associate professor in the Department of Computer Science and Engin eering at the University of Notre Dame. His research is on computational m odels for learning human languages\, particularly how to translate from on e language to another. His work on applying formal grammars and machine le arning to translation has been recognized with two best paper awards (at A CL 2005 and NAACL HLT 2009). He has received research grants from DARPA\, NSF\, Google\, and Amazon\, has served on the executive board of NAACL and the editorial board of Computational Linguistics and JAIR\, and is curren tly on the editorial board of Transactions of the ACL. DTSTART;TZID=America/New_York:20221017T120000 DTEND;TZID=America/New_York:20221017T131500 LOCATION:Hackerman Hall B17 @ 3400 N. Charles Street\, Baltimore\, MD 21218 SEQUENCE:0 SUMMARY:David Chiang (University of Notre Dame) “Exact Recursive Probabilis tic Programming with Colin McDonald\, Darcey Riley\, Kenneth Sible (Notre Dame) and Chung-chieh Shan (Indiana)” URL:https://www.clsp.jhu.edu/events/david-chiang-university-of-notre-dame/ X-COST-TYPE:free X-ALT-DESC;FMTTYPE=text/html:\\n\\n\\n
Abstr act
\n