Search Results
Tag: LLMClear filters
Running MiniMax M2.5 Locally on NVIDIA DGX Spark
How I got a 230B parameter open model running on desktop hardware using NVIDIA DGX Spark, Unsloth quantization, and llama.cpp - matching cloud API performance without cloud dependencies.
Running MiniMax M2.5 Locally with Claude Code
A quick how-to guide for connecting Claude Code to your local MiniMax M2.5 inference server.
Meet Reachy Mini: Building an AI-Powered Conference Badge Reader
How I built a fun conference booth experience combining an open-source robot, vision AI, and Python. Plus: exploring local LLMs as the next step.
Llama and DeepSeek with LibreChat for Conversational AI
Step-by-step guide: Deploy Llama and DeepSeek LLMs using SGLang on DataCrunch, and integrate with LibreChat for seamless conversations.
What are Large Language Models and Key Terminologies
Michael Mueller6 min read
Learn about Large Language Models (LLMs), the Transformer architecture, tokenization, embeddings, and their impact on natural language processing.