How to inject knowledge efficiently? Knowledge infusion scaling law for LLMs

Why do LLMs freak out over the seahorse emoji?

Reducing Context Bloat with Dynamic Context Loading (DCL) for LLMs & MCP

LLMs are the ultimate demoware

Which table format do LLMs understand best?

Ask HN: Estimation of copyright material used by LLM

Create diverse responses from single prompt to LLMs using Beam search

Every LLM Is Its Own Media Channel

Let's Build the GPT Tokenizer: A Complete Guide to Tokenization in LLMs

Show HN: Git for LLMs – A context management interface

Show HN: Incremental JSON parser for streaming LLM tool calls in Ruby

Git considers SHA-256, Rust, LLMs, and more

Made a repo to gather and generate wrong tech info that can affect LLM poisoning — could be used as a counter-dataset too.

De-emojifying scripts - setting yourself apart from LLMs

Nvidia DGX Spark and Apple Mac Studio = 4x Faster LLM Inference with EXO 1.0

Show HN: I built an open-source AI data layer that connects any LLM to any data

'Western Qwen': IBM Wows with Granite 4 LLM Launch and Hybrid Mamba/Transformer

valuetier.org (and some thoughts on LLMs)

Mercury: Unlocking Multi-GPU Operator Optimization for LLMs via Remote Memory Scheduling

Seeking Advice on Intent Recognition Architecture: Keyword + LLM Fallback, Context Memory, and Prompt Management

New Generation Bio-inspired AI Architecture: Moving Beyond LLM Statistical Models

Chezmoi introduces ban on LLM-generated contributions

A beginner's guide to deploying LLMs with AMD on Windows using PyTorch

A Simple Way to Explore Codebases with LLMs

🏛️ Building LLMs from Scratch – Part 2: Data Collection & Custom Tokenizers

Agon: A Terminal UI for Comparing Local LLMs with Ollama

ProofOfThought: LLM-based reasoning using Z3 theorem proving

A Go toolkit for evaluating and scoring LLM/AI responses

Efficient LLM:Bandwidth, Compute, Synchronization, and Capacity are all you need

I created and open sourced an LLM and backend orchestration system

More →