António Branco’s research while affiliated with University of Lisbon and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (128)


Figure 1: Top: Architecture of the method presented in this paper. Bottom: Three examples, showing inputs (left images and edit requests) from the MAGICBRSUH test set and outputs (right images) from our proposed method.
Figure 3: Examples of different edit-direction weights, with base images and instructions from MAGICBRUSH. Instruction in top row: "Make the woman obese."; in bottom row: "Let's add birds to the sky".
Figure 7: Performance scores with CLIP-I of the best prompt at each meta-prompting step (blue line), conducted on the first 100 examples of the test set, and the manual prompt used in the first set of experiments (red line)
Hands-off Image Editing: Language-guided Editing without any Task-specific Labeling, Masking or even Training
  • Preprint
  • File available

February 2025

·

12 Reads

Rodrigo Santos

·

António Branco

·

·

Instruction-guided image editing consists in taking an image and an instruction and deliverring that image altered according to that instruction. State-of-the-art approaches to this task suffer from the typical scaling up and domain adaptation hindrances related to supervision as they eventually resort to some kind of task-specific labelling, masking or training. We propose a novel approach that does without any such task-specific supervision and offers thus a better potential for improvement. Its assessment demonstrates that it is highly effective, achieving very competitive performance.

Download




From greatest simplicity to full power

October 2024

·

13 Reads

Language Resources and Evaluation

While language processing services are key assets for the science and technology of language, the possible ways under which they may be made available to the widest range of their end users are critical to support an Open Science policy for this scientific domain. Although providing such processing services under some web-based interface, at large, offers itself as an immediate and cogent response to that challenge, turning this view into an effective access to language processing services is an undertaking deserving a clear conceptual direction and a corresponding robust empirical validation. Based on an extensive overview of major undertakings towards making language processing tools available and on the design principles worked out and implemented in the PORTULAN CLARIN infrastructure, in this paper we advocate for a Research-Infrastructure-as-a-Service (RIaaS) model. This model unleashes accessibility to language processing services in as many web-based interface modalities as the current stage of technological development permits to support, in order to serve as many types of end users as possible, from IT developers to Digital Humanities researchers, and including citizen scientists, teachers, students and digital artists among many others.


Sentence encoders not specifically for Portuguese. Second column reports sizes of models (parameters) and embeddings (bits).
Results on STS, with Spearman correlation coefficient.
Open Sentence Embeddings for Portuguese with the Serafim PT* encoders family

July 2024

·

27 Reads

Sentence encoder encode the semantics of their input, enabling key downstream applications such as classification, clustering, or retrieval. In this paper, we present Serafim PT*, a family of open-source sentence encoders for Portuguese with various sizes, suited to different hardware/compute budgets. Each model exhibits state-of-the-art performance and is made openly available under a permissive license, allowing its use for both commercial and research purposes. Besides the sentence encoders, this paper contributes a systematic study and lessons learned concerning the selection criteria of learning objectives and parameters that support top-performing encoders.


Meta-prompting Optimized Retrieval-augmented Generation

July 2024

·

146 Reads

Retrieval-augmented generation resorts to content retrieved from external sources in order to leverage the performance of large language models in downstream tasks. The excessive volume of retrieved content, the possible dispersion of its parts, or their out of focus range may happen nevertheless to eventually have a detrimental rather than an incremental effect. To mitigate this issue and improve retrieval-augmented generation, we propose a method to refine the retrieved content before it is included in the prompt by resorting to meta-prompting optimization. Put to empirical test with the demanding multi-hop question answering task from the StrategyQA dataset, the evaluation results indicate that this method outperforms a similar retrieval-augmented system but without this method by over 30%.


Advancing Neural Encoding of Portuguese with Transformer Albertina PT-*

December 2023

·

93 Reads

·

30 Citations

Lecture Notes in Computer Science

To advance the neural encoding of Portuguese (PT), and a fortiori the technological preparation of this language for the digital age, we developed a Transformer-based foundation model that sets a new state of the art in this respect for two of its variants, namely European Portuguese from Portugal (PT-PT) and American Portuguese from Brazil (PT-BR). To develop this encoder, which we named Albertina PT-*, a strong model was used as a starting point, DeBERTa, and its pre-training was done over data sets of Portuguese, namely over a data set we gathered for PT-PT and over the brWaC corpus for PT-BR. The performance of Albertina and competing models was assessed by evaluating them on prominent downstream language processing tasks adapted for Portuguese. Both Albertina versions are distributed free of charge and under a most permissive license possible and can be run on consumer-grade hardware, thus seeking to contribute to the advancement of research and innovation in language technology for Portuguese.


WebLicht overview
The CLARIN infrastructure as an interoperable language technology platform for SSH and beyond

June 2023

·

96 Reads

·

7 Citations

Language Resources and Evaluation

CLARIN is a European Research Infrastructure Consortium developing and providing a federated and interoperable platform to support scientists in the field of the Social Sciences and Humanities in carrying-out language-related research. This contribution provides an overview of the entire infrastructure with a particular focus on tool interoperability, ease of access to research data, tools and services, the importance of sharing knowledge within and across (national) communities, and community building. By taking into account FAIR principles from the very beginning, CLARIN succeeded in becoming a successful example of a research infrastructure that is actively used by its members. The benefits CLARIN members reap from their infrastructure secure a future for their common good that is both sustainable and attractive to partners beyond the original target groups.



Citations (61)


... InstructPix2Pix [11] fine-tuned diffusion models on synthetic image editing pairs guided by instructions. More recent models [5,12,4,13,14,15,16,17,18,19,20,21,22,23,24] have advanced the semantic and compositional understanding of edits. ...

Reference:

GIE-Bench: Towards Grounded Evaluation for Text-Guided Image Editing
Leveraging LLMs for On-the-Fly Instruction Guided Image Editing
  • Citing Chapter
  • November 2024

... One solution would be to create LMs specific to a certain language variety. However, this presents its own set of challenges for training and evaluation (Armengol-Estapé et al. 2021;Rodrigues et al. 2023). Training a LM, be it from scratch or fine-tuning, requires a large, carefully curated corpus for training and several benchmarks for evaluation (Albalak et al. 2024), which are either nonexistent or contaminated by the dominant language varieties. ...

Advancing Neural Encoding of Portuguese with Transformer Albertina PT-*
  • Citing Chapter
  • December 2023

Lecture Notes in Computer Science

... The results reported here were partially supported by PORTULAN CLA-RIN-Research Infrastructure for the Science and Technology of Language, funded by Lisboa2020, Alentejo2020 and FCT-Fundação para a Ciência e Tecnologia under the Grant PINFRA/22117/2016. The PORTULAN CLARIN Workbench comprises a number of tools that are based on a large body of research work contributed by different authors and teams, which continues to grow and is acknowledged here: (Barreto et al., 2006;Branco et al., 2010Branco et al., , 2011Branco et al., , 2012Branco et al., , 2014Cruz et al., 2018;Veiga et al., 2011;Branco & Henriques, 2003;Silva et al., 2009;Rodrigues et al., 2016Rodrigues et al., , 2020Costa & Branco, 2012;Santos et al., 2019;Miranda et al., 2011). ...

Automatic syllabification of Portuguese

Revista da Associação Portuguesa de Linguística

... To enhance the uptake, visibility, operation, and transnational collaboration of K-centres, CLARIN developed a variety of initiatives. Their list and detailed descriptions can be found on the CLARIN website under the tile "Learn & Exchange" 5 as well as in overview articles by Jong et al. (2022), Jong et al. (2018), and Branco et al. (2023) among others. We will briefly present some of initiatives grouping them according to their functions. ...

The CLARIN infrastructure as an interoperable language technology platform for SSH and beyond

Language Resources and Evaluation

... Regarding Portuguese language, Rodrigues et al. [9] introduced Albertina, a Transformer-based foundation model both for European (Portugal) and American (Brazil) Portuguese variants. Compared to BERTimbau, it uses 128-token sequence truncation instead of 512. ...

Advancing Neural Encoding of Portuguese with Transformer Albertina PT-*

... Results show that current systems fall short of human explanation quality. Branco et al. (2021) raised fundamental concerns about the reliability of commonsense benchmarks, showing that models exploit shallow dataset artifacts instead of reasoning. Another paper presents ATOMIC 2020, a comprehensive commonsense knowledge graph that enhances the reasoning capabilities of AI models (Hwangy et al. ...

Shortcutted Commonsense: Data Spuriousness in Deep Learning of Commonsense Reasoning
  • Citing Conference Paper
  • January 2021

... Afterward, users are offered a list of other members who fit these criteria. Members can choose the people they want to meet via chat, email, or phone, by either sending a personal letter or simply relying on reinterpretations in a more public letter, on a bulletin board [14]. Language exchange platforms consist of three modes of interaction. ...

The European language technology landscape in 2020 : language-centric and human-centric AI for cross-cultural communication in multilingual Europe

... Posteriormente, surgiram abordagens de aprendizagem profunda baseadas em embeddings de palavras pré-treinados, como os gerados usando Word2Vec (Mikolov et al., 2013) (e.g., Cha et al., 2017;Nadeem & Ostendorf, 2018;Filighera et al., 2019). Por fim, mais recentemente, a investigação naárea mudouse para o ajuste de modelos pré-treinados baseados em transformadores, como BERT (Devlin RoBERTa (Liu et al., 2019); por exemplo, (Santos et al., 2021;Yancey et al., 2021;Martinc et al., 2021;Mohtaj et al., 2022). ...

Neural Text Categorization with Transformers for Learning Portuguese as a Second Language
  • Citing Chapter
  • September 2021

Lecture Notes in Computer Science

... Statistical significance To assess the statistical significance of the improvements by our method, we employed the unpaired t-test. 3 We evaluated the baseline system with three seeds and did the same for the meta-prompting optimized system. Both samples are independent and one may assume the samples are normally distributed. ...

Comparative Probing of Lexical Semantics Theories for Cognitive Plausibility and Technological Usefulness
  • Citing Conference Paper
  • January 2020

... A tradução pode ser realizada por humanos ou com a assistência de ferramentas de tradução automática. Para o português, exemplos desta abordagem incluem a tradução de referências conhecidas para avaliar a similaridade semântica e analogia (Querido et al., 2017); ou a tradução do conjunto de dados SQuAD para resposta a perguntas (Carvalho et al., 2021). ...

LX-LR4DistSemEval: a collection of language resources for the evaluation of distributional semantic models of Portuguese

Revista da Associação Portuguesa de Linguística