Search is not available for this dataset
title
string | arxiv_id
string |
---|---|
SPEECH: Structured Prediction with Energy-Based Event-Centric Hyperspheres | 2305.13617 |
Rule By Example: Harnessing Logical Rules for Explainable Hate Speech Detection | 2307.12935 |
Pruning Pre-trained Language Models Without Fine-Tuning | 2210.06210 |
When Does Translation Require Context? A Data-driven, Multilingual Exploration | 2109.07446 |
Do Androids Laugh at Electric Sheep? Humor “Understanding” Benchmarks from The New Yorker Caption Contest | 2209.06293 |
DiffusionDB: A Large-scale Prompt Gallery Dataset for Text-to-Image Generative Models | 2210.14896 |
Being Right for Whose Right Reasons? | 2306.00639 |
Glot500: Scaling Multilingual Corpora and Language Models to 500 Languages | 2305.12182 |
A Theory of Unsupervised Speech Recognition | 2306.07926 |
DIONYSUS: A Pre-trained Model for Low-Resource Dialogue Summarization | 2212.10018 |
Diverse Demonstrations Improve In-context Compositional Generalization | 2212.06800 |
Self-Adaptive In-Context Learning: An Information Compression Perspective for In-Context Example Selection and Ordering | 2212.10375 |
On the Efficacy of Sampling Adapters | 2307.03749 |
Measuring Progress in Fine-grained Vision-and-Language Understanding | 2305.07558 |
Elaboration-Generating Commonsense Question Answering at Scale | 2209.01232 |
White-Box Multi-Objective Adversarial Attack on Dialogue Generation | 2305.03655 |
Few-shot Adaptation Works with UnpredicTable Data | 2208.01009 |
In-Context Analogical Reasoning with Pre-Trained Language Models | 2305.17626 |
Peek Across: Improving Multi-Document Modeling via Cross-Document Question-Answering | 2305.15387 |
REV: Information-Theoretic Evaluation of Free-Text Rationales | 2210.04982 |
Schema-Guided User Satisfaction Modeling for Task-Oriented Dialogues | 2305.16798 |
SimLM: Pre-training with Representation Bottleneck for Dense Passage Retrieval | 2207.02578 |
From Ultra-Fine to Fine: Fine-tuning Ultra-Fine Entity Typing Models to Fine-grained | 2312.06188 |
What Makes Pre-trained Language Models Better Zero-shot Learners? | 2209.15206 |
Cross2StrA: Unpaired Cross-lingual Image Captioning with Cross-lingual Cross-modal Structure-pivoted Alignment | 2305.12260 |
Plan-and-Solve Prompting: Improving Zero-Shot Chain-of-Thought Reasoning by Large Language Models | 2305.04091 |
Symbolic Chain-of-Thought Distillation: Small Models Can Also “Think” Step-by-Step | 2306.14050 |
Generating EDU Extracts for Plan-Guided Summary Re-Ranking | 2305.17779 |
Gradient-based Intra-attention Pruning on Pre-trained Language Models | 2212.07634 |
DiffusEmp: A Diffusion Model-Based Framework with Multi-Grained Control for Empathetic Response Generation | 2306.01657 |
Summary-Oriented Vision Modeling for Multimodal Abstractive Summarization | 2212.07672 |
InfoMetIC: An Informative Metric for Reference-free Image Caption Evaluation | 2305.06002 |
HistRED: A Historical Document-Level Relation Extraction Dataset | 2307.04285 |
PVGRU: Generating Diverse and Relevant Dialogue Responses via Pseudo-Variational Mechanism | 2212.09086 |
A Survey on Zero Pronoun Translation | 2305.10196 |
MPCHAT: Towards Multimodal Persona-Grounded Conversation | 2305.17388 |
Dual-Alignment Pre-training for Cross-lingual Sentence Embedding | 2305.09148 |
Alleviating Over-smoothing for Unsupervised Sentence Representation | 2305.06154 |
From Characters to Words: Hierarchical Pre-trained Language Model for Open-vocabulary Language Understanding | 2305.14571 |
Code4Struct: Code Generation for Few-Shot Event Structure Prediction | 2210.12810 |
Efficient Semiring-Weighted Earley Parsing | 2307.02982 |
Entity Tracking in Language Models | 2305.02363 |
WACO: Word-Aligned Contrastive Learning for Speech Translation | 2212.09359 |
Knowledge-enhanced Mixed-initiative Dialogue System for Emotional Support Conversations | 2305.10172 |
Parameter-Efficient Fine-Tuning without Introducing New Latency | 2305.16742 |
MASSIVE: A 1M-Example Multilingual Natural Language Understanding Dataset with 51 Typologically-Diverse Languages | 2204.08582 |
DiffusionBERT: Improving Generative Masked Language Models with Diffusion Models | 2211.15029 |
Unified Demonstration Retriever for In-Context Learning | 2305.04320 |
DimonGen: Diversified Generative Commonsense Reasoning for Explaining Concept Relationships | 2212.10545 |
Hidden Schema Networks | 2207.03777 |
Towards Robust Low-Resource Fine-Tuning with Multi-View Compressed Representations | 2211.08794 |
Pre-Training to Learn in Context | 2305.09137 |
Privacy-Preserving Domain Adaptation of Semantic Parsers | 2212.10520 |
KILM: Knowledge Injection into Encoder-Decoder Language Models | 2302.09170 |
Tokenization and the Noiseless Channel | 2306.16842 |
Reasoning with Language Model Prompting: A Survey | 2212.09597 |
DISCO: Distilling Counterfactuals with Large Language Models | 2212.10534 |
SCOTT: Self-Consistent Chain-of-Thought Distillation | 2305.01879 |
Evaluating Open-Domain Question Answering in the Era of Large Language Models | 2305.06984 |
What the DAAM: Interpreting Stable Diffusion Using Cross Attention | 2210.04885 |
Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training | 2206.00621 |
Counterspeeches up my sleeve! Intent Distribution Learning and Persistent Fusion for Intent-Conditioned Counterspeech Generation | 2305.13776 |
What is the best recipe for character-level encoder-only modelling? | 2305.05461 |
Dialect-robust Evaluation of Generated Text | 2211.00922 |
TOME: A Two-stage Approach for Model-based Retrieval | 2305.11161 |
miCSE: Mutual Information Contrastive Learning for Low-shot Sentence Embeddings | 2211.04928 |
Forgotten Knowledge: Examining the Citational Amnesia in NLP | 2305.18554 |
Measuring the Instability of Fine-Tuning | 2302.07778 |
Multilingual LLMs are Better Cross-lingual In-context Learners with Alignment | 2305.05940 |
Long-Tailed Question Answering in an Open World | 2305.06557 |
Did the Models Understand Documents? Benchmarking Models for Language Understanding in Document-Level Relation Extraction | 2306.11386 |
ContraCLM: Contrastive Learning For Causal Language Model | 2210.01185 |
Prompting Language Models for Linguistic Structure | 2211.07830 |
FLamE: Few-shot Learning from Natural Language Explanations | 2306.08042 |
Fact-Checking Complex Claims with Program-Guided Reasoning | 2305.12744 |
Patton: Language Model Pretraining on Text-Rich Networks | 2305.12268 |
Soft Language Clustering for Multilingual Model Pre-training | 2306.07610 |
Dynamic Transformers Provide a False Sense of Efficiency | 2305.12228 |
Multi-target Backdoor Attacks for Code Pre-trained Models | 2306.08350 |
Multi-Level Knowledge Distillation for Out-of-Distribution Detection in Text | 2211.11300 |
MMDialog: A Large-scale Multi-turn Dialogue Dataset Towards Multi-modal Open-domain Conversation | 2211.05719 |
ByGPT5: End-to-End Style-conditioned Poetry Generation with Token-free Language Models | 2212.10474 |
Matching Pairs: Attributing Fine-Tuned Models to their Pre-Trained Large Language Models | 2306.09308 |
Large Language Models Meet NL2Code: A Survey | 2212.09420 |
One Network, Many Masks: Towards More Parameter-Efficient Transfer Learning | 2305.17682 |
WebIE: Faithful and Robust Information Extraction on the Web | 2305.14293 |
Contextual Knowledge Learning for Dialogue Generation | 2305.18200 |
Text Style Transfer Back-Translation | 2306.01318 |
Continual Knowledge Distillation for Neural Machine Translation | 2212.09097 |
CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding | 2209.10918 |
Few-Shot Document-Level Event Argument Extraction | 2209.02203 |
Towards Understanding and Improving Knowledge Distillation for Neural Machine Translation | 2305.08096 |
DSEE: Dually Sparsity-embedded Efficient Tuning of Pre-trained Language Models | 2111.00160 |
Do CoNLL-2003 Named Entity Taggers Still Work Well in 2023? | 2212.09747 |
Weakly Supervised Vision-and-Language Pre-training with Relative Representations | 2305.15483 |
RECAP: Retrieval-Enhanced Context-Aware Prefix Encoder for Personalized Dialogue Response Generation | 2306.07206 |
On “Scientific Debt” in NLP: A Case for More Rigour in Language Model Pre-Training Research | 2306.02870 |
Element-aware Summarization with Large Language Models: Expert-aligned Evaluation and Chain-of-Thought Method | 2305.13412 |
What does the Failure to Reason with “Respectively” in Zero/Few-Shot Settings Tell Us about Language Models? | 2305.19597 |
FormNetV2: Multimodal Graph Contrastive Learning for Form Document Information Extraction | 2305.02549 |
End of preview. Expand
in Dataset Viewer.
README.md exists but content is empty.
- Downloads last month
- 39