Category Added in a WPeMatico Campaign
Language models (LMs) face significant challenges related to privacy and copyright concerns due to their training on vast amounts of text data. The inadvertent inclusion of private and copyrighted content in training datasets has led to legal and ethical issues, including copyright lawsuits and compliance requirements with regulations like GDPR. Data owners increasingly demand the…
As LLMs become increasingly integral to various AI tasks, their massive parameter sizes lead to high memory requirements and bandwidth consumption. While quantization-aware training (QAT) offers a potential solution by allowing models to operate with lower-bit representations, existing methods often require extensive training resources, making them impractical for large models. The research paper addresses the…
Evaluating large language models (LLMs) has become increasingly challenging due to their complexity and versatility. Ensuring the reliability and quality of these models’ outputs is crucial for advancing AI technologies and applications. Researchers need help developing reliable evaluation methods to assess the accuracy and impartiality of LLMs’ outputs, given human evaluations’ subjective, inconsistent, and costly…
High-dimensional clinical data (HDCD) refers to datasets in healthcare where the number of variables (or features) is significantly larger than the number of patients (or observations). As the number of variables increases, the data space grows exponentially, requiring substantial computational resources that make it difficult to process and analyze. Additionally, models built on high-dimensional data…
Large language models (LLMs) have showcased remarkable capabilities in generating content and solving complex problems across various domains. However, a notable challenge persists in their ability to perform multi-step deductive reasoning. This type of reasoning requires a coherent and logical thought process over extended interactions, which current LLMs need help with due to their training…
Using offline web apps and AI apps often comes with challenges. Users typically need to navigate multiple steps to get an app running. These steps can be confusing and time-consuming, especially for those who are not tech-savvy. Additionally, managing and customizing these apps often requires manual editing of files, making the process even more cumbersome.…
Evaluating the retrieval and reasoning capabilities of large language models (LLMs) in extremely long contexts, extending up to 1 million tokens, is a significant challenge. Efficiently processing long texts is crucial for extracting relevant information and making accurate decisions based on extensive data. This challenge is particularly relevant for real-world applications, such as legal document…
Despite their expanding capabilities, large language models (LLMs) need help with processing extensive contexts. These limitations stem from Transformer-based architectures struggling to extrapolate beyond their training window size. Processing long token sequences requires substantial computational resources and risks producing noisy attention embeddings. These constraints hinder LLMs’ ability to incorporate domain-specific, private, or up-to-date information effectively.…
Innovation and the artistic, musical, and literary expression of human experiences and emotions depend on creativity. However, the idea that material created by humans is inherently better is coming under pressure from the emergence of generative artificial intelligence (AI) technologies, such as Large Language Models (LLMs). Content in several formats, such as text (ChatGPT), graphics…
LLMs excel in natural language processing tasks but face deployment challenges due to high computational and memory demands during inference. Recent research [MWM+24, WMD+23, SXZ+24, XGZC23, LKM23] aims to enhance LLM efficiency through quantization, pruning, distillation, and improved decoding. Sparsity, a key approach, reduces computation by omitting zero elements and lessens I/O transfer between memory…