Jules White’s research while affiliated with Vanderbilt University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (195)


FIGURE EE Processing time comparison by prompt style (averaged across datasets).
Enhancing structured data generation with GPT-4o evaluating prompt efficiency across prompt styles
  • Article
  • Full-text available

March 2025

·

25 Reads

Frontiers in Artificial Intelligence

Ashraf Elnashar

·

Jules White

·

Large language models (LLMs), such as GPT-4o, provide versatile techniques for generating and formatting structured data. However, prompt style plays a critical role in determining the accuracy, efficiency, and token cost of the generated outputs. This paper explores the effectiveness of three specific prompt styles–JSON, YAML, and Hybrid CSV/Prefix–for structured data generation across diverse applications. We focus on scenarios such as personal stories, receipts, and medical records, using randomized datasets to evaluate each prompt style's impact. Our analysis examines these prompt styles across three key metrics: accuracy in preserving data attributes, token cost associated with output generation, and processing time required for completion. By incorporating structured validation and comparative analysis, we ensure precise evaluation of each prompt style's performance. Results are visualized through metrics-based comparisons, such as Prompt Style vs. Accuracy, Prompt Style vs. Token Cost, and Prompt Style vs. Processing Time. Our findings reveal trade-offs between prompt style complexity and performance, with JSON providing high accuracy for complex data, YAML offering a balance between readability and efficiency, and Hybrid CSV/Prefix excelling in token and time efficiency for flat data structures. This paper explores the pros and cons of applying the GPT-4o LLM to generate structured data. It also provides practical recommendations for selecting prompt styles tailored to specific requirements, such as data integrity, cost-effectiveness, and real-time processing needs. Our findings contribute to research on how prompt engineering can optimize structured data generation for AI-driven applications, as well as documenting limitations that motivate future work needed to improve LLMs for complex tasks.

Download

Building Living Software Systems with Generative & Agentic AI

August 2024

·

49 Reads

This paper is an opinion paper that looks at the future of computing in the age of Generative \& Agentic AI. Current software systems are static and inflexible, leading to significant challenges in translating human goals into computational actions. "Living software systems" powered by generative AI offer a solution to this fundamental problem in computing. Traditional software development involves multiple layers of imperfect translation, from business requirements to code, resulting in rigid systems that struggle to adapt to changing user needs and contexts. Generative AI, particularly large language models, can serve as a universal translator between human intent and computer operations. This approach enables the creation of more flexible, context-aware systems that can dynamically evolve to meet user goals. Two pathways for implementing living software systems are explored: using generative AI to accelerate traditional software development, and leveraging agentic AI to create truly adaptive systems. New skills like Prompt Engineering are necessary. By reimagining software as a living, adaptable entity, we can create computing interfaces that are more intuitive, powerful, and responsive to human needs.



Evaluating Persona Prompting for Question Answering Tasks

June 2024

·

38 Reads

·

8 Citations

Using large language models (LLMs) effectively by applying prompt engineering is a timely research topic due to the advent of highly performant LLMs, such as ChatGPT-4. Various patterns of prompting have proven effective, including chain-of-thought, self-consistency, and personas. This paper makes two contributions to research on prompting patterns. First, we measure the effect of single- and multi-agent personas in various knowledge-testing, multiple choice, and short answer environments, using a variation of question answering tasks known as as ”openness.” Second, we empirically evaluate several persona-based prompting styles on 4,000+ questions. Our results indicate that single-agent expert personas perform better on high-openness tasks and that effective prompt engineering becomes more important for complex multi-agent methods.


Towards a Catalog of Prompt Patterns to Enhance the Discipline of Prompt Engineering

June 2024

·

82 Reads

·

19 Citations

ACM SIGAda Ada Letters

The rapid advent of Large Language Models (LLMs), such as ChatGPT and Claude, is revolutionizing various fields, from education and healthcare to the engineering of reliable software systems. These LLMs operate through "prompts," which are natural language inputs that users employ to query and leverage the models' capabilities. Given the novelty of LLMs, the understanding of how to effectively use prompts remains largely anecdotal, based on isolated use cases. This fragmented approach limits the reliability and utility of LLMs, especially when they are applied in mission-critical software environments. To harness the full potential of LLMs in such crucial contexts, therefore, we need a systematic, disciplined approach to "prompt engineering" that guides interactions with and evaluations of these LLMs.


ChatGPT Prompt Patterns for Improving Code Quality, Refactoring, Requirements Elicitation, and Software Design

June 2024

·

84 Reads

·

198 Citations

This chapter presents design techniques for software engineering, in the form of prompt patterns, to solve common problems that arise when using large language models (LLMs) to automate common software engineering activities, such as ensuring code is decoupled from third-party libraries and creating API specifications from lists of requirements. This chapter provides two contributions to research on using LLMs for software engineering. First, it provides a catalog of patterns for software engineering that classifies patterns according to the types of problems they solve. Second, it explores several prompt patterns that have been applied to improve requirements elicitation, rapid prototyping, code quality, deployment, and testing.






Citations (75)


... They focus on comparing the multiple solutions suggested by Copilot. Doderlein et al. [15] and Elnashar et al. [16] analyze the impact of prompt engineering on the runtime performance of solutions generated by Copilot and ChatGPT respectively. Mastropaolo et al. [17] analyze the solutions generated by Copilot for different semantically equivalent task descriptions. ...

Reference:

Measuring the Runtime Performance of Code Produced with GitHub Copilot
Evaluating the Performance of LLM-Generated Code for ChatGPT-4 and AutoGen Along with Top-Rated Human Solutions
  • Citing Conference Paper
  • January 2024

... At the end of the study we also explored the use of generative AI to describe model code. We used the persona pattern (Olea et al., 2024) from prompt engineering and asked the AI tool to take the role of an elective surgery patient. We provided a prompt including all the CCU model code and asked for a description of potential journeys through the model. ...

Evaluating Persona Prompting for Question Answering Tasks
  • Citing Conference Paper
  • June 2024

... Ένας τρόπος να εκμεταλλευτεί κανείς στο μέγιστο δυνατό τις δυνατότητες των LLM, όπως το ChatGPT είναι μέσω του prompt engineering, που είναι μια διαδικασία σχεδιασμού και βελτίωσης των προτροπών που παρέχονται σε ένα LLΜ με στόχο την πρόκληση ποιοτικότερων και πιο στοχευμένων απαντήσεων εκ μέρους των LLM (Schmidt et al., 2024). Σε ένα εκπαιδευτικό πλαίσιο, το prompt engineering μπορεί να βελτιώσει την εκπαιδευτική εμπειρία από την μεριά του εκπαιδευόμενου προσαρμόζοντας κατάλληλα τις απαντήσεις των LLM στις εξειδικευμένες εκπαιδευτικές ανάγκες της κάθε τάξης και του κάθε μαθητή. ...

Towards a Catalog of Prompt Patterns to Enhance the Discipline of Prompt Engineering
  • Citing Article
  • June 2024

ACM SIGAda Ada Letters

... Recent research on applying LLMs in software engineering, particularly in processing requirements, has shown promise. For example, a study by White et al. [14] shows how LLMs like ChatGPT improve software engineering tasks such as generating API specifications from requirements lists. Another study by Belzner et al. [15] highlights how LLMs can assist in generating class design and corresponding pseudocode of the important classes and their relationships. ...

ChatGPT Prompt Patterns for Improving Code Quality, Refactoring, Requirements Elicitation, and Software Design
  • Citing Chapter
  • June 2024

... Large Language Models (LLMs) are underpinned by advanced deep learning frameworks, with transformers serving as their foundational architecture. Transformers, introduced by Vaswani [73,74] revolutionized natural language processing (NLP) by addressing the limitations of recurrent and convolutional neural networks through self-attention mechanisms and parallel processing. Unlike traditional models, transformers process entire input sequences simultaneously, which enhances computational efficiency and enables the capture of long-range dependencies essential for understanding complex language patterns [73,74] (Figure 4). ...

Semantic Compression with Large Language Models
  • Citing Conference Paper
  • November 2023

... Full-text assessment for eligibility resulted in 55 articles which met the review criteria ( Figure 1) and underwent data extraction as well as quality and risk of bias appraisal. There were 13 articles on ataxia [7][8][9][10][11][12][13][14][15][16][17][18][19], 11 articles on chorea [20][21][22][23][24][25][26][27][28][29][30], five articles on dystonia [31][32][33][34][35], one article on tics [36], and 25 articles on tremor . No articles on myoclonus were included. ...

A Multi-stage Transfer Learning Strategy for Diagnosing a Class of Rare Laryngeal Movement Disorders
  • Citing Article
  • November 2023

Computers in Biology and Medicine

... Given the limited size of our dataset, we believed that allocating a portion to a validation set could adversely impact the model's performance. Moreover, research indicated that with small datasets, the models often perform best with default hyperparameters, and that hyperparameter tuning might negatively influence performance (14,15). These factors led us to the decision of not engaging in hyperparameter tuning and adpoting a k-fold Cross Validation with k = 10. ...

Large Language Models to generate meaningful feature model instances

... Ethereum, a decentralized computing system utilizing the Ethereum Virtual Machine (EVM) and native token ETH, enables the creation and execution of Smart Contracts, facilitating the storage and codification of data operations. Ethereum's payment policy, based on 'gas,' insentience's network nodes to validate and execute transactions while deterring malicious attacks [18]. DLT serves diverse purposes in the medical field, providing a secure and immutable ledger for storing EHRs and patient data. ...

Design pattern recommendations for building decentralized healthcare applications

Frontiers in Blockchain

... A pressing question is how to move beyond ad-hoc prompting toward reusable, maintainable prompts that can be systematically applied and iteratively improved. Recent research indicates that well-structured prompts can significantly enhance LLM performance and reliability, making it possible to integrate these models into complex workflows (Singh et al., 2023;White et al., 2023). This effort aligns with "humanity-centered automation," where technology should serve human values, be transparent, and respect ethical constraints (Reiff-Stephan, 2024 In this work, we propose six key concepts of reusable prompts: Versioning, Model Selection, Purpose Definition, Variables, Examples, and Output Structuring. ...

ChatGPT Prompt Patterns for Improving Code Quality, Refactoring, Requirements Elicitation, and Software Design

... This is a proof-of-concept study. Drawing on existing guidance on prompt engineering [29,30] and our experience in building human-AI dialogues [20,31], we developed a series of flipped interactions through a series of iterative refinements. In this approach, the AI initiates the conversation and prompts the human expert to provide the necessary data, enabling a collaborative process for generating and refining item templates. ...

A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT