Author: Divyesh Vitthal Jawkhede

Divyesh Vitthal Jawkhede
66 POSTS0 COMMENTS
Divyesh is a consulting intern at Marktechpost. He is pursuing a BTech in Agricultural and Food Engineering from the Indian Institute of Technology, Kharagpur. He is a Data Science and Machine learning enthusiast who wants to integrate these leading technologies into the agricultural domain and solve challenges.

Align-Pro: A Cost-Effective Alternative to RLHF for LLM Alignment

Aligning large language models (LLMs) with human values is essential as these models become central to various societal functions. A significant challenge arises when...

Step Towards Best Practices for Open Datasets for LLM Training

Large language models rely heavily on open datasets to train, which poses significant legal, technical, and ethical challenges in managing such datasets. There are...

Meet OmAgent: A New Python Library for Building Multimodal Language Agents

Understanding long videos, such as 24-hour CCTV footage or full-length films, is a major challenge in video processing. Large Language Models (LLMs) have shown...

MinMo: A Multimodal Large Language Model with Approximately 8B Parameters for Seamless Voice Interaction

Advances in large language and multimodal speech-text models have laid a foundation for seamless, real-time, natural, and human-like voice interactions. Achieving this requires systems...

TimeDP: A Multi-Domain Time Series Diffusion Model with Domain Prompts

Generating time series data is important for many applications, including data augmentation, synthetic datasets, and scenarios. However, when there is more than one, this...

Meet Search-o1: An AI Framework that Integrates the Agentic Search Workflow into the o1-like Reasoning Process of LRM for Achieving Autonomous Knowledge Supplementation

Large reasoning models are developed to solve difficult problems by breaking them down into smaller, manageable steps and solving each step individually. The models...

Can LLMs Design Good Questions Based on Context? This AI Paper Evaluates Questions Generated by LLMs from Context, Comparing Them to Human-Generated Questions

Large Language Models (LLMs) are used to create questions based on given facts or context, but understanding how good these questions are can be...

From Contradictions to Coherence: Logical Alignment in AI Models

Large Language Models (LLMs) aim to align with human preferences, ensuring reliable and trustworthy decision-making. However, these models acquire biases, logical leaps, and hallucinations,...

Unlocking Cloud Efficiency: Optimized NUMA Resource Mapping for Virtualized Environments

Disaggregated systems are a new type of architecture designed to meet the high resource demands of modern applications like social networking, search, and in-memory...

From Latent Spaces to State-of-the-Art: The Journey of LightningDiT

Latent diffusion models are advanced techniques for generating high-resolution images by compressing visual data into a latent space using visual tokenizers. These tokenizers reduce...

University of South Florida Researchers Propose TeLU Activation Function for Fast and Stable Deep Learning

Inspired by the brain, neural networks are essential for recognizing images and processing language. These networks rely on activation functions, which enable them to...

Mixture-of-Denoising Experts (MoDE): A Novel Generalist MoE-based Diffusion Policy

Diffusion Policies in Imitation Learning (IL) can generate diverse agent behaviors, but as models grow in size and capability, their computational demands increase, leading...