view article Article Efficient LLM Pretraining: Packed Sequences and Masked Attention Oct 7, 2024 β’ 70
Running 3.79k The Ultra-Scale Playbook π 3.79k The ultimate guide to training LLM on large GPU Clusters
Running on L4 Agents 1.19k ControlNet V1.1 π 1.19k Generate edited images using edge, pose, and other guides
Running Agents Featured 998 ControlNet π 998 Generate images from sketches, edges, poses, and depth maps
Build error 253 Controlnet for Interior Design π₯ 253 Upload an image and edit it using segmentation, inpainting, or regeneration
LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders Paper β’ 2404.05961 β’ Published Apr 9, 2024 β’ 66
Linear Transformers with Learnable Kernel Functions are Better In-Context Models Paper β’ 2402.10644 β’ Published Feb 16, 2024 β’ 81