-
Prompt Cache: Modular Attention Reuse for Low-Latency Inference
Paper • 2311.04934 • Published • 32 -
I2VEdit: First-Frame-Guided Video Editing via Image-to-Video Diffusion Models
Paper • 2405.16537 • Published • 17 -
NV-Embed: Improved Techniques for Training LLMs as Generalist Embedding Models
Paper • 2405.17428 • Published • 19
Pedro Batista
pedrovhb
·
AI & ML interests
None yet
Recent Activity
liked
a Space
4 days ago
linoyts/FLUX.2-klein-relight-brush
liked
a Space
8 months ago
nvidia/PartPacker
liked
a model
8 months ago
nanonets/Nanonets-OCR-s
Organizations
None yet