Pagine meno recenti
Vengono mostrati sotto 50 risultati dal 201 al 250.
- Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets (10:10, 19 ago 2024)
- Grokking (11:25, 19 ago 2024)
- Are We Done with MMLU? (11:30, 19 ago 2024)
- Alice in Wonderland: Simple Tasks Showing Complete Reasoning Breakdown in State-Of-the-Art Large Language Models (11:53, 19 ago 2024)
- Confabulazione (12:00, 19 ago 2024)
- Language Models are Few-Shot Learners (13:09, 19 ago 2024)
- How Many Data Points is a Prompt Worth? (13:10, 19 ago 2024)
- Prefix-Tuning (13:17, 19 ago 2024)
- Prefix-tuning: Optimizing continuous prompts for generation (13:18, 19 ago 2024)
- LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models (09:44, 22 ago 2024)
- ConceptNet (20:36, 24 ago 2024)
- What's in a Name? -- Gender Classification of Names with Character Based Machine Learning Models (20:45, 26 ago 2024)
- End-to-End Object Detection with Transformers (11:53, 28 ago 2024)
- Generating Sequences With Recurrent Neural Networks (2014) (11:54, 28 ago 2024)
- Detection Transformer (11:56, 28 ago 2024)
- Non-Maximum Suppression (12:43, 28 ago 2024)
- YOLO-World: Real-Time Open-Vocabulary Object Detection (13:02, 28 ago 2024)
- Judging LLM-as-a-judge with MT-Bench and Chatbot Arena (20:11, 28 ago 2024)
- LLM-as-a-judge (20:12, 28 ago 2024)
- Vocabolario (10:46, 5 set 2024)
- None (10:20, 6 set 2024)
- Deep Contextualized Word Representations (10:24, 6 set 2024)
- A Neural Probabilistic Language Model (10:27, 6 set 2024)
- Neural Machine Translation by Jointly Learning to Align and Translate (10:29, 6 set 2024)
- Hallucinating Faces (11:02, 6 set 2024)
- Allucinazione (11:11, 6 set 2024)
- Red Teaming (11:16, 6 set 2024)
- Chain of Thought Prompting Elicits Reasoning in Large Language Models (11:26, 6 set 2024)
- Survey of Hallucination in Natural Language Generation (11:29, 6 set 2024)
- Chain of Thought (11:30, 6 set 2024)
- Functional Benchmarks for Robust Evaluation of Reasoning Performance, and the Reasoning Gap (12:00, 6 set 2024)
- Efficient Estimation of Word Representations in Vector Space (12:01, 6 set 2024)
- Going Deeper with Convolutions (12:10, 6 set 2024)
- Pop Music Transformer: Beat-based Modeling and Generation of Expressive Pop Piano Compositions (13:44, 6 set 2024)
- Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank (18:27, 6 set 2024)
- Convolutional Neural Networks for Sentence Classification (18:54, 6 set 2024)
- LLaMA: Open and Efficient Foundation Language Models (19:02, 6 set 2024)
- Masked-Language-Modeling (MLM) (19:18, 6 set 2024)
- BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension (06:30, 8 set 2024)
- BART (06:32, 8 set 2024)
- Generative Pretrained Transformer (GPT) (06:32, 8 set 2024)
- Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation (18:08, 8 set 2024)
- Forward pass (18:10, 8 set 2024)
- XLSTM: Extended Long Short-Term Memory (04:53, 9 set 2024)
- Long Short-Term Memory (LSTM) (04:55, 9 set 2024)
- Long Short-Term Memory (04:56, 9 set 2024)
- BLEU: a method for automatic evaluation of machine translation (05:01, 9 set 2024)
- Bilingual Evaluation Understudy (BLEU) (05:02, 9 set 2024)
- Obiettivo di pre-training (10:37, 10 set 2024)
- What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization? (10:40, 10 set 2024)