stereoplegic
's Collections
Embeddings
updated
Towards General Text Embeddings with Multi-stage Contrastive Learning
Paper
•
2308.03281
•
Published
•
1
NEFTune: Noisy Embeddings Improve Instruction Finetuning
Paper
•
2310.05914
•
Published
•
14
EELBERT: Tiny Models through Dynamic Embeddings
Paper
•
2310.20144
•
Published
•
3
Dynamic Word Embeddings for Evolving Semantic Discovery
Paper
•
1703.00607
•
Published
•
1
Paper
•
1702.08359
•
Published
•
1
Unleashing the Power of Pre-trained Language Models for Offline
Reinforcement Learning
Paper
•
2310.20587
•
Published
•
16
Understanding and Improving Information Transfer in Multi-Task Learning
Paper
•
2005.00944
•
Published
•
1
Split, Encode and Aggregate for Long Code Search
Paper
•
2208.11271
•
Published
•
1
TEAL: Tokenize and Embed ALL for Multi-modal Large Language Models
Paper
•
2311.04589
•
Published
•
18
Frustratingly Simple Memory Efficiency for Pre-trained Language Models
via Dynamic Embedding Pruning
Paper
•
2309.08708
•
Published
•
3
CokeBERT: Contextual Knowledge Selection and Embedding towards Enhanced
Pre-Trained Language Models
Paper
•
2009.13964
•
Published
•
1
Plug-and-Play Knowledge Injection for Pre-trained Language Models
Paper
•
2305.17691
•
Published
•
1
Plug-and-Play Document Modules for Pre-trained Models
Paper
•
2305.17660
•
Published
•
1
Identifying Linear Relational Concepts in Large Language Models
Paper
•
2311.08968
•
Published
•
1
Can the Inference Logic of Large Language Models be Disentangled into
Symbolic Concepts?
Paper
•
2304.01083
•
Published
•
1
Sub-Sentence Encoder: Contrastive Learning of Propositional Semantic
Representations
Paper
•
2311.04335
•
Published
•
1
Retrieve Anything To Augment Large Language Models
Paper
•
2310.07554
•
Published
•
6
Nomic Embed: Training a Reproducible Long Context Text Embedder
Paper
•
2402.01613
•
Published
•
14
Improving Text Embeddings with Large Language Models
Paper
•
2401.00368
•
Published
•
79
Uncovering hidden geometry in Transformers via disentangling position
and context
Paper
•
2310.04861
•
Published
Char2Subword: Extending the Subword Embedding Space Using Robust
Character Compositionality
Paper
•
2010.12730
•
Published
CharBERT: Character-aware Pre-trained Language Model
Paper
•
2011.01513
•
Published
Training Multilingual Pre-trained Language Model with Byte-level
Subwords
Paper
•
2101.09469
•
Published
Neural Machine Translation with Byte-Level Subwords
Paper
•
1909.03341
•
Published
byteSteady: Fast Classification Using Byte-Level n-Gram Embeddings
Paper
•
2106.13302
•
Published
MAGNET: Improving the Multilingual Fairness of Language Models with
Adaptive Gradient-Based Tokenization
Paper
•
2407.08818
•
Published
OFA: A Framework of Initializing Unseen Subword Embeddings for Efficient
Large-scale Multilingual Continued Pretraining
Paper
•
2311.08849
•
Published
•
5