Large language models (LLMs) have introduced impressive capabilities, particularly in reasoning tasks. Models like OpenAI’s O1 utilize “long-thought reasoning,” where complex problems are broken into manageable steps and solutions are refined iteratively. While this approach enhances problem-solving, it comes at a cost: extended output sequences lead to increased computational time and energy use. These inefficiencies…
Smartphones are essential tools in dAIly life. However, the complexity of tasks on mobile devices often leads to frustration and inefficiency. Navigating applications and managing multi-step processes consumes time and effort. Advancements in AI have introduced large multimodal models (LMMs) that enable mobile assistants to perform intricate operations autonomously. While these innovations aim to simplify…
The study of autonomous agents powered by large language models (LLMs) has shown great promise in enhancing human productivity. These agents are designed to assist in various tasks such as coding, data analysis, and web navigation. They allow users to focus on creative and strategic work by automating routine digital tasks. However, despite the advancements,…
The advent of advanced AI models has led to innovations in how machines process information, interact with humans, and execute tasks in real-world settings. Two emerging pioneering approaches are large concept models (LCMs) and large action models (LAMs). While both extend the foundational capabilities of large language models (LLMs), their objectives and applications diverge. LCMs…
Aligning large language models (LLMs) with human values is essential as these models become central to various societal functions. A significant challenge arises when model parameters cannot be updated directly because the models are fixed or inaccessible. In these cases, the focus is on adjusting the input prompts to make the model’s outputs match the…
Evaluating conversational AI systems powered by large language models (LLMs) presents a critical challenge in artificial intelligence. These systems must handle multi-turn dialogues, integrate domain-specific tools, and adhere to complex policy constraints—capabilities that traditional evaluation methods struggle to assess. Existing benchmarks rely on small-scale, manually curated datasets with coarse metrics, failing to capture the dynamic…
Proteins, essential macromolecules for biological processes like metabolism and immune response, follow the sequence-structure-function paradigm, where amino acid sequences determine 3D structures and functions. Computational protein science AIms to decode this relationship and design proteins with desired properties. Traditional AI models have achieved significant success in specific protein modeling tasks, such as structure prediction and…
Pre-trained vision models have been foundational to modern-day computer vision advances across various domains, such as image classification, object detection, and image segmentation. There is a rather massive amount of data inflow, creating dynamic data environments that require a continual learning process for our models. New regulations for data privacy require specific information to be…
A fundamental challenge in advancing AI research lies in developing systems that can autonomously perform structured reasoning and dynamically expand domain knowledge. Traditional AI models often rely on implicit reasoning processes, which limit their ability to explain decisions, adapt across domains, and generalize relational patterns. These shortcomings hinder their applicability to complex scientific problems that…