Prompt Engineering

227 papers with code • 16 benchmarks • 16 datasets

Prompt engineering is the process of designing and refining the prompts used to generate text from language models, such as GPT-3 or similar models. The goal of prompt engineering is to improve the quality and relevance of the generated text by carefully crafting the prompts to elicit the desired responses from the model.

Prompt engineering involves several steps, including selecting the appropriate model architecture and parameters, designing the prompt format and structure, selecting the appropriate task and training data, and fine-tuning the model using the selected prompt and data.

Prompt engineering is a crucial step in the development of language models, as it can greatly influence the quality and effectiveness of the model's responses. By carefully designing and refining the prompts used to generate text, researchers and developers can improve the accuracy and relevance of the model's output, making it more useful for a wide range of applications, including chatbots, language translation, content creation, and more.

Libraries

Use these libraries to find Prompt Engineering models and implementations

Easy Problems That LLMs Get Wrong

autogenai/easy-problems-that-llms-get-wrong 30 May 2024

We introduce a comprehensive Linguistic Benchmark designed to evaluate the limitations of Large Language Models (LLMs) in domains such as logical reasoning, spatial intelligence, and linguistic understanding, among others.

4
30 May 2024

ORLM: Training Large Language Models for Optimization Modeling

cardinal-operations/orlm 28 May 2024

We apply the data from OR-Instruct to various open-source LLMs of 7b size (termed as ORLMs), resulting in a significantly improved capability for optimization modeling.

6
28 May 2024

Optimizing Large Language Models for OpenAPI Code Completion

BohdanPetryshyn/code-llama-fim-fine-tuning 24 May 2024

Recent advancements in Large Language Models (LLMs) and their utilization in code generation tasks have significantly reshaped the field of software development.

1
24 May 2024

Detection and Positive Reconstruction of Cognitive Distortion sentences: Mandarin Dataset and Evaluation

405200144/Dataset-of-Cognitive-Distortion-detection-and-Positive-Reconstruction 24 May 2024

In this study, we emphasize the theoretical foundation for the Positive Reconstruction Framework, grounded in broaden-and-build theory.

0
24 May 2024

A Lost Opportunity for Vision-Language Models: A Comparative Study of Online Test-time Adaptation for Vision-Language Models

mariodoebler/test-time-adaptation 23 May 2024

Through a systematic exploration of prompt-based techniques and existing test-time adaptation methods, the study aims to enhance the adaptability and robustness of vision-language models in diverse real-world scenarios.

144
23 May 2024

E2TP: Element to Tuple Prompting Improves Aspect Sentiment Tuple Prediction

mghiasvandm/E2TP 10 May 2024

Generative approaches have significantly influenced Aspect-Based Sentiment Analysis (ABSA), garnering considerable attention.

2
10 May 2024

Exploring the Capabilities of Large Multimodal Models on Dense Text

yuliang-liu/multimodalocr 9 May 2024

To further explore the capabilities of LMM in complex text tasks, we propose the DT-VQA dataset, with 170k question-answer pairs.

335
09 May 2024

Can We Use Large Language Models to Fill Relevance Judgment Holes?

chuanmeng/qpp-genre 9 May 2024

Incomplete relevance judgments limit the re-usability of test collections.

10
09 May 2024

CityLLaVA: Efficient Fine-Tuning for VLMs in City Scenario

alibaba/aicity2024_track2_aliopentrek_cityllava 6 May 2024

In the vast and dynamic landscape of urban settings, Traffic Safety Description and Analysis plays a pivotal role in applications ranging from insurance inspection to accident prevention.

25
06 May 2024

Towards A Human-in-the-Loop LLM Approach to Collaborative Discourse Analysis

oele-isis-vanderbilt/AIED24_LBR 6 May 2024

LLMs have demonstrated proficiency in contextualizing their outputs using human input, often matching or beating human-level performance on a variety of tasks.

0
06 May 2024