Loading...

Tag trends are in beta. Feedback? Thoughts? Email me at [email protected]

Should LLMs just treat text content as an image?

The IP risks of LLMs at work are nontrivial

How do LLM's trade off lives between different categories?

OptPipe: Memory- and Scheduling-Optimized Pipeline Parallelism for LLM Training

Fast-DLLM: Training-Free Acceleration of Diffusion LLM

LLMs can get "brain rot"

Identifying Life-Changing Books with LLMs (2024)

Two things LLM coding agents are still bad at

Karpathy on DeepSeek-OCR paper: Are pixels better inputs to LLMs than text?

The security paradox of local LLMs

LaTeX, LLMs and Boring Technology

We tested 20 LLMs for ideological bias, revealing distinct alignments

Writing an LLM from scratch, part 22 – training our LLM

A small number of samples can poison LLMs of any size

What are some ongoing topics in Computer Science research that don't involve AI/ML (and definitely LLMs)?

ChunkLLM: A Lightweight Pluggable Framework for Accelerating LLMs Inference

My trick for getting consistent classification from LLMs

Neural audio codecs: how to get audio into LLMs

LLMs are getting better at character-level text manipulation

Show HN: Create-LLM – Train your own LLM in 60 seconds

LLMs are mortally terrified of exceptions

Don't Force Your LLM to Write Terse [Q/Kdb] Code: An Information Theory Argument

AdapTive-LeArning Speculator System (ATLAS): Faster LLM inference

Show HN: LLM Rescuer – Fixing the billion dollar mistake in Ruby

Game over. AGI is not imminent, and LLMs are not the royal road to getting there

Reasoning LLMs are wandering solution explorers

OpenAI Cofounder Builds New Open Source LLM 'Nanochat' - and Doesn't Use Vibe Coding

Writing an LLM from scratch, part 20 – starting training, and cross entropy loss

Building a High-Performance LLM Gateway in Go: Bifrost (50x Faster than LiteLLM)

Building LLM inference libraries in pure Java and run them with LangChain4j locally on GPUs (No CUDA, No C++)

More →