The last installment of our RAG series compares RAG vs. fine tuning alternatives. The engineering edition looks at OpenAI's new agentic APIs. research section dives into Microsoft's Phi-4 new models. In our opinion essay we will debate another controversial topic. Small foundation models is one of the most fascinating trends in generative AI. Seeing how relatively small models can match the capabilities of mega models is truly amazing. Last week we had two remarkable releases in this area: Command A and Gemma 3. Command A, developed by Cohere, is engineered to match or surpass the performance of leading models like GPT-4o and DeepSeek-V3 across various enterprise tasks. Notably, it achieves this efficiency while operating on just two GPUs, a stark contrast to other models that may require up to 32 GPUs. This reduction in hardware dependency translates to faster processing speeds'Command A processes information at a rate of 156 tokens per second, outpacing GPT-4o by 1.75 times and...
learn more