Open post test-time compute

Test-Time Compute: The Next Frontier in AI Scaling

Major AI labs, including OpenAI, are shifting their focus away from building ever-larger language models (LLMs). Instead, they are exploring "test-time compute", where models receive extra processing time during execution to produce better results. This change stems from the limitations of traditional pre-training methods, which have reached a plateau in performance and are becoming too...

Open post smaller

The Trend Towards Smaller Language Models in AI

The landscape of artificial intelligence (AI) is undergoing a notable transformation, shifting from the pursuit of ever-larger language models (LLMs) to the development of smaller, more efficient models. This shift, driven by technological advancements and practical considerations, is redefining how AI systems are built, deployed, and utilized across various sectors. The Shift in AI Model...

Open post tokens

Understanding Tokens in AI

In the realm of Artificial Intelligence (AI), particularly in the context of Large Language Models (LLMs), tokens are the fundamental units of language that enable machines to comprehend and generate human-like text. This article delves into the world of tokens, exploring their definition, history, types, and significance in the AI landscape. What are Tokens? In...

Open post malicious

FraudGPT and other malicious AIs are the new frontier of online threats. What can we do?

The internet, a vast and indispensable resource for modern society, has a darker side where malicious activities thrive. From identity theft to sophisticated malware attacks, cyber criminals keep coming up with new scam methods. Widely available generative artificial intelligence (AI) tools have now added a new layer of complexity to the cyber security landscape. Staying...

Open post philosophy

Why Software Engineers Should Embrace Philosophy

In the world of software engineering, where lines of code and caffeine reign supreme, there’s an unlikely ally waiting in the wings: philosophy. Yes, you read that right. The ancient art of pondering existence, ethics, and the meaning of life can turbocharge your coding skills. Here’s why software engineers should dive into the realm of...

Open post LoRA and Finetuning

LoRA vs. Fine-Tuning LLMs

LoRA (Low-Rank Adaptation) and fine-tuning are two methods to adapt large language models (LLMs) to specific tasks or domains. LLMs are pre-trained on massive amounts of general domain data, such as GPT-3, RoBERTa, and DeBERTa, and have shown impressive performance on various natural language processing (NLP) tasks. Why fine tune a LLM? Fine-tuning of LLMs...

Open post Super Apps vs Large Language Models

Super Apps vs. Large Language Models

In the rapidly evolving world of software, Super Apps and Large Language Models (LLMs) offer two distinct approaches to address user needs and preferences. Super Apps, like WeChat, consolidate various functions and services into a single platform, while LLMs, such as GPT from OpenAI promise a new era of software adaptability through self-modifying code generation....

Scroll to top