Author: Mohammad Asjad

Mohammad Asjad
256 POSTS0 COMMENTS
Asjad is an intern consultant at Marktechpost. He is persuing B.Tech in mechanical engineering at the Indian Institute of Technology, Kharagpur. Asjad is a Machine learning and deep learning enthusiast who is always researching the applications of machine learning in healthcare.

DeltaProduct: An AI Method that Balances Expressivity and Efficiency of the Recurrence Computation, Improving State-Tracking in Linear Recurrent Neural Networks

The Transformer architecture revolutionised natural language processing with its self-attention mechanism, enabling parallel computation and effective context retrieval. However, Transformers face significant limitations when...

PydanticAI: Advancing Generative AI Agent Development through Intelligent Framework Design

Innovative frameworks that simplify complex interactions with large language models have fundamentally transformed the landscape of generative AI development in Python. PydanticAI emerges as...

TxAgent: An AI Agent that Delivers Evidence-Grounded Treatment Recommendations by Combining Multi-Step Reasoning with Real-Time Biomedical Tool Integration

Precision therapy has emerged as a critical approach in healthcare, tailoring treatments to individual patient profiles to optimise outcomes while reducing risks. However, determining...

Building a Retrieval-Augmented Generation (RAG) System with FAISS and Open-Source LLMs

Retrieval-augmented generation (RAG) has emerged as a powerful paradigm for enhancing the capabilities of large language models (LLMs). By combining LLMs' creative generation abilities...

Meet PC-Agent: A Hierarchical Multi-Agent Collaboration Framework for Complex Task Automation on PC

Multi-modal Large Language Models (MLLMs) have demonstrated remarkable capabilities across various domains, propelling their evolution into multi-modal agents for human assistance. GUI automation agents...

Implementing Text-to-Speech TTS with BARK Using Hugging Face’s Transformers library in a Google Colab environment

Text-to-Speech (TTS) technology has evolved dramatically in recent years, from robotic-sounding voices to highly natural speech synthesis. BARK is an impressive open-source TTS model...

Salesforce AI Releases Text2Data: A Training Framework for Low-Resource Data Generation

Generative AI faces a critical challenge in balancing autonomy and controllability. While autonomy has advanced significantly through powerful generative models, controllability has become a...

Q-Filters: A Training-Free AI Method for Efficient KV Cache Compression

Large Language Models (LLMs) have significantly advanced due to the Transformer architecture, with recent models like Gemini-Pro1.5, Claude-3, GPT4, and Llama3.1 demonstrating capabilities to...

Starter Guide For Running Large Language Models LLMs

Running large language models (LLMs) presents significant challenges due to their hardware demands, but numerous options exist to make these powerful tools accessible. Today's...

Thinking Harder, Not Longer: Evaluating Reasoning Efficiency in Advanced Language Models

Large language models (LLMs) have progressed beyond basic natural language processing to tackle complex problem-solving tasks. While scaling model size, data, and compute has...

CoSyn: An AI Framework that Leverages the Coding Capabilities of Text-only Large Language Models (LLMs) to Automatically Create Synthetic Text-Rich Multimodal Data

Vision-language models (VLMs) have demonstrated impressive capabilities in general image understanding, but face significant challenges when processing text-rich visual content such as charts, documents,...

Why Do Task Vectors Exist in Pretrained LLMs? This AI Research from MIT and Improbable AI Uncovers How Transformers Form Internal Abstractions and the...

Large Language Models (LLMs) have demonstrated remarkable similarities to human cognitive processes' ability to form abstractions and adapt to new situations. Just as humans...