Deepmind AI Research Foundations Part 3
Notes I took whilst studying "Google DeepMind: AI Research Foundations". Covers fine tuning and accelerating your model.

Notes I took whilst studying "Google DeepMind: AI Research Foundations". Covers fine tuning and accelerating your model.

Notes I took whilst studying "Google DeepMind: AI Research Foundations". Covers designing and training neural networks as well as the transformer architecture.

Notes I took whilst studying "Google DeepMind: AI Research Foundations". Covers building your own small language model and language representation.

Notes on architecting AI inference stacks and TPUs from Google's learning path, "Inference on TPUs".

Notes on architecting multi-agent systems from Google's learning path, "Architect Multi-Agent Systems with Agent Development Kit".

Notes and ideas on annotations and LLMs. Using annotations in conjunction with LLM dev tooling as well as generating annotation processors with LLMs

I never write about caches and caching, so I thought I'd cover some basics on LLM caching. Covers inference and prompt caching.

Notes on the TensorZero LLM gateway. Covers templates, schemas, feedback, retries, evals, DICL, MIPRO, model-prompt-inference optimization.

Some basics on Ollama. Includes some details on quantization, vector DBs, model storage, model format and modelfiles.

Comparisons of the OpenAI service offering with that of Anthropic. Includes context window, rate limits and model optimization.
