S. R. K. Branavan’s research while affiliated with Massachusetts Institute of Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (13)


Learning Document-Level Semantic Properties from Free-Text Annotations.
  • Article

January 2009

·

53 Reads

·

87 Citations

Journal of Artificial Intelligence Research

S. R. K. Branavan

·

Harr Chen

·

Jacob Eisenstein

·

Regina Barzilay

This paper presents a new method for inferring the semantic properties of documents by lever- aging free-text keyphrase annotations. Such annotations are becoming increasingly abundant due to the recent dramatic growth in semi-structured, user-generated online content. One especially relevant domain is product reviews, which are often annotated by their authors with pros/cons keyphrases such as "a real bargain" or "good value." These annotations are representative of the underlying semantic properties; however, unlike expert annotations, they are noisy: lay authors may use different labels to denote the same property, and some labels may be missing. To learn using such noisy annotations, we find a hidden paraphrase structure which clusters the keyphrases. The paraphrase structure is linked with a latent topic model of the review texts, enabling the sys- tem to predict the properties of unannotated documents and to effectively aggregate the semantic properties of multiple reviews. Our approach is implemented as a hierarchical Bayesian model with joint inference. We find that joint inference increases the robustness of the keyphrase clustering and encourages the latent topics to correlate with semantically meaningful properties. Multiple evalua- tions demonstrate that our model substantially outperforms alternative approaches for summarizing single and multiple documents into a set of semantically salient keyphrases.


Learning Document-Level Semantic Properties from Free-Text Annotations

January 2008

·

35 Reads

·

64 Citations

Journal of Artificial Intelligence Research

This paper presents a new method for inferring the semantic properties of documents by leveraging free-text keyphrase annotations. Such annotations are becoming increasingly abundant due to the recent dramatic growth in semi-structured, user-generated online content. One especially relevant domain is product reviews, which are often annotated by their authors with pros/cons keyphrases such as "a real bargain" or "good value." These annotations are representative of the underlying semantic properties; however, unlike expert annotations, they are noisy: lay authors may use different labels to denote the same property, and some labels may be missing. To learn using such noisy annotations, we find a hidden paraphrase structure which clusters the keyphrases. The paraphrase structure is linked with a latent topic model of the review texts, enabling the system to predict the properties of unannotated documents and to effectively aggregate the semantic properties of multiple reviews. Our approach is implemented as a hierarchical Bayesian model with joint inference. We find that joint inference increases the robustness of the keyphrase clustering and encourages the latent topics to correlate with semantically meaningful properties. Multiple evaluations demonstrate that our model substantially outperforms alternative approaches for summarizing single and multiple documents into a set of semantically salient keyphrases.


Generating a Table-of-Contents.
  • Conference Paper
  • Full-text available

January 2007

·

34 Reads

·

10 Citations

This paper presents a method for the auto- matic generation of a table-of-contents. This type of summary could serve as an effec- tive navigation tool for accessing informa- tion in long texts, such as books. To gen- erate a coherent table-of-contents, we need to capture both global dependencies across different titles in the table and local con- straints within sections. Our algorithm ef- fectively handles these complex dependen- cies by factoring the model into local and global components, and incrementally con- structing the model's output. The results of automatic evaluation and manual assessment confirm the benefits of this design: our sys- tem is consistently ranked higher than non- hierarchical baselines.

Download

Citations (12)


... For the non-NLP (multi-modal) tasks, most focused on environment-grounded language learning, i.e., driving the agent to associate natural language instructions with the environments and make corresponding reactions, such as selecting mentioned objects from an image/video (Matuszek et al. 2012;Krishnamurthy and Kollar 2013;Puig et al. 2018), following navigational instructions to move the agent (Tellex et al. 2011;Kim and Mooney 2012;Chen 2012;Artzi and Zettlemoyer 2013;Bisk, Yuret, and Marcu 2016), plotting corresponding traces on a map (Vogel and Jurafsky 2010; Chen and Mooney 2011), playing soccer/card games based on given rules (Kuhlmann et al. 2004;Eisenstein et al. 2009;Branavan, Silver, and Barzilay 2011;Babeş-Vroman et al. 2012;Goldwasser and Roth 2011), generating real-time sports broadcast (Chen and Mooney 2008;Liang, Jordan, and Klein 2009), controlling software (Branavan, Zettlemoyer, and Barzilay 2010), and querying external databases (Clarke et al. 2010), among others. Meanwhile, instructions are also widely adapted to help communicate with the system in solving NLP tasks, for example, following instructions to manipulate strings (Gaddy and Klein 2019), classifying e-mails based on the given explanations Mitchell 2017, 2018), and text-to-code generation (Acquaviva et al. 2022). ...

Reference:

Large Language Model Instruction Following: A Survey of Progresses and Challenges
Learning to Win by Reading Manuals in a Monte-Carlo Framework
  • Citing Conference Paper
  • January 2011

... Understanding action preconditions and effects in text is a crucial yet challenging task. Branavan et al. (2012) pioneer work in this area using reinforcement learning to extract high-level planning knowledge from text with the guidance of action preconditions and effects. Dalvi et al. (2018) develop a dataset and models for paragraph comprehension, and highlight the importance of tracking state changes in procedural text. ...

Learning High-Level Planning from Text
  • Citing Conference Paper
  • July 2012

... Another straightforward approach to using natural language for reward shaping can be seen in [26], which describes rules using natural language to decide whether to choose an action. [27][28] also perform language-to-reward mapping by designing rewards based on if the agent arrives at the right location or not by following the instruction provided. These approaches still face the problem of efficiently defining reward functions since it requires an expert programmer who can make decisions about how language instructions are mapped to the environment. ...

Learning to Win by Reading Manuals in a Monte-Carlo Framework
  • Citing Article
  • January 2014

Journal of Artificial Intelligence Research

... A few sentiment analysis systems take as input the sample speech of the user and perform prosodic processing of the speech signals. [16] takes the speech signal as input, converts it into text and then, only considers the text content for the affect analysis. The feature extraction is performed on the text review and adjective-noun pairs are identified as potential sentiment phrases. ...

Good grief, I can speak it! preliminary experiments in audio restaurant reviews

·

·

S.R.K. Branavan

·

[...]

·

Regina Barzilay

... One of the most well studied class of models on the space of rankings/permutations are the celebrated Mallows models, first introduced in [24]. Since then, Mallows models (and their variants) have received significant attention in statistics ( [9,10,13,14,15,25,28]), probability ( [3,6,12,17,19,29,30]), and machine learning ( [2,8,21,22,23,26,27]). In [28], the authors introduce a class of exponential family models on the space of permutations, which includes some of the commonly studied Mallows models. ...

Content Modeling Using Latent Permutations
  • Citing Article
  • October 2009

Journal of Artificial Intelligence Research

... Before the prevalence of Large Language Model, traditional autonomous agents primarily implement through reinforcement learning (Branavan et al., 2009;Shvo et al., 2021;Gur et al., 2022), semantic parsing (Li et al., 2020) and imitation learning (Humphreys et al., 2022) that clones human's keyboard and mouse actions . The recent trend is to use Large Language Model to generate GUI instructions and actions. ...

Reinforcement Learning for Mapping Instructions to Actions
  • Citing Conference Paper
  • January 2009

... Fine-tuning Virtual Agents traditionally relies on humanannotated datasets, which are labor-intensive and timeconsuming . Methods such as imitation learning (Humphreys et al., 2022) and reinforcement learning (Branavan et al., 2009;2010) have been employed to fine-tune agents based on curated expert trajectories or outcome rewards, but these approaches often suffer from compounding errors and limited exploration (Christianos et al., 2023;Xi et al., 2024;Song et al., 2024). Recent advancements, such as reject sampling fine-tuning (RFT) (Yuan et al., 2023) and direct policy optimization (DPO) (Rafailov et al., 2023), have sought to reduce reliance on human annotations by leveraging both successful and failure trajectories (Lai et al., 2024;Zhang et al., 2024c). ...

Reading Between the Lines: Learning to Map High-level Instructions to Commands
  • Citing Conference Paper
  • December 2010

... The originality of the work stems from using the n-gram concept to model nodes in a tree in addition to modeling sequential entities. Branavan, Deshpande and Barzilay (2007) propose a method that automatically generates a table-of-contents structure for long documents such as books. They first segment a document hierarchically and then generate an informative title for each segment. ...

Generating a Table-of-Contents.

... Mais ces modèles ne sont pas adéquats dans le cas où le domaine etudié possède une multitude d'aspects dont certains apparaissent trés rarement dans le corpus. Ainsi, certains auteurs [6,10,12,18] ont plutôt adopté des méthodes basées sur la modélisation des sujets. ...

Learning Document-Level Semantic Properties from Free-Text Annotations
  • Citing Conference Paper
  • January 2008

Journal of Artificial Intelligence Research

... Text semantic segmentation necessitates that sentences within the same paragraph cohesively revolve around a central topic, while maintaining minimal semantic overlap between distinct paragraphs. Early unsupervised methodologies identified segmentation boundaries through Bayesian models (Chen et al., 2009;Riedl and Biemann, 2012) or graph-based methods, where sentences were treated as nodes (Glavaš et al., 2016). On the other hand, supervised methods have leveraged pretrained language models (PLMs) derived from extensive corpora, subsequently fine-tuning them on annotated text semantic segmentation datasets. ...

Global Models of Document Structure using Latent Permutations.
  • Citing Conference Paper
  • January 2009