RunPod Blog
  • RunPod
  • Docs
Sign in Subscribe

Language Models

A collection of 3 posts
No-Code AI: How I Ran My First Language Model Without Coding
No-Code AI

No-Code AI: How I Ran My First Language Model Without Coding

I wanted to run an open-source AI model myself—no code, just curiosity. Here’s how I deployed Mistral 7B on a cloud GPU and what I learned.
03 Apr 2025 8 min read
How Much VRAM Does Your LLM Need? A Guide to GPU Memory Requirements
GPU Power

How Much VRAM Does Your LLM Need? A Guide to GPU Memory Requirements

Discover how to determine the right VRAM for your Large Language Model (LLM). Learn about GPU memory requirements, model parameters, and tools to optimize your AI deployments.
08 Jul 2024 5 min read
Run Larger LLMs on RunPod Serverless Than Ever Before - Llama-3 70B (and beyond!)
Language Models

Run Larger LLMs on RunPod Serverless Than Ever Before - Llama-3 70B (and beyond!)

Up until now, RunPod has only supported using a single GPU in Serverless, with the exception of using two 48GB cards (which honestly didn't help, given the overhead involved in multi-GPU setups for LLMs.) You were effectively limited to what you could fit in 80GB, so you would
06 Jun 2024 3 min read
Page 1 of 1
RunPod Blog © 2025
  • Sign up
Powered by Ghost