Tag: LLM

  • Test Cases for Training LLMs

    Test Cases for Training LLMs When training Large Language Models (LLMs), particularly for tasks like **extracting information from tax documents**, writing effective test cases is crucial for ensuring your model learns as intended and can accurately perform the desired function. These test cases differ significantly from traditional software testing due to the probabilistic and generative Read more

  • Implementing Locally running Mistral Chatbot with RAG

    Locally running Mistral Chatbot with RAG Let’s implement a local running chatbot with Mistral LLM using RAG to retrieve documents from a locally running Vector DB that also contains FAQs. Here’s a breakdown of the steps and the Python code to achieve this: Phase 1: Setting Up the Local Environment Install Dependencies: pip install transformers Read more

  • Top 10 LLMs on Hugging Face for Chatbot & RAG Use (Early May 2025)

    Top 10 LLMs on Hugging Face for Chatbot & RAG This list is based on a combination of factors including general popularity, instruction-following capabilities, context window size, and community interest relevant to chatbot and Retrieval-Augmented Generation (RAG) applications. 1. mistralai/Mixtral-8x7B-Instruct-v0.1 Use Cases: Excellent for instruction following, complex reasoning in chatbots, and can handle long contexts Read more

  • Top 10 LLMs on Hugging Face & Use Cases: Part 2

    Another Top 10 LLMs on Hugging Face & Use Cases Another Top 10 Popular LLMs on Hugging Face & Their Use Cases Here’s another selection of popular and interesting Large Language Models available on Hugging Face, showcasing the diversity of the open-source LLM landscape as of early May 2025. 1. google/gemma-7b-it Use Cases: Instruction tuning, Read more

  • Top 10 LLMs on Hugging Face & Use Cases

    Top 10 LLMs on Hugging Face & Use Cases Top 10 Popular LLMs on Hugging Face & Their Use Cases Please note that “top” can be subjective and based on various factors like downloads, recent interest, and performance on specific benchmarks. This list reflects a mix of widely used and influential models as of early Read more

  • Using local LLM for Document Extraction

    Non-Cloud LLM for Document Extraction This guide explains how to use a non-cloud version of a pretrained Large Language Model (LLM) for document extraction, focusing on open-source models and local execution. Phase 1: Setting Up Your Local Environment 1. Hardware Requirements Ensure your system meets the following recommendations: CPU/GPU: An NVIDIA GPU with sufficient VRAM Read more

  • CUDA vs. ROCm for LLM Training

    CUDA vs. ROCm CUDA (Compute Unified Device Architecture) and ROCm (Radeon Open Compute) are the two primary software platforms for General-Purpose computing on Graphics Processing Units (GPGPU) used in accelerating computationally intensive tasks, including the training of Large Language Models (LLMs). CUDA is developed by NVIDIA and is designed for their GPUs, while ROCm is Read more

  • Exploring CUDA (Compute Unified Device Architecture)

    Exploring CUDA CUDA is a parallel computing platform and programming model developed by NVIDIA for use with their GPUs. It allows software developers to leverage the massive parallel processing power of NVIDIA GPUs for general-purpose computing tasks, significantly accelerating applications beyond traditional CPU-bound processing. 1. CUDA Architecture: The Hardware Foundation NVIDIA GPUs are designed with Read more

  • Can AMD GPUs Train LLMs?

    Can AMD GPUs Train LLMs? AMD GPUs can be used to train Large Language Models (LLMs). While NVIDIA GPUs, particularly those with CUDA architecture, have historically dominated the LLM training landscape, AMD has been making significant strides in this area with its ROCm (Radeon Open Compute) platform. 1. ROCm Platform ROCm is AMD’s open-source software Read more

  • AMD GPUs vs. NVIDIA GPUs for LLM Training

    AMD GPUs vs. NVIDIA GPUs for LLM Training Here we dive into how AMD GPUs can be used for LLM training, and compare them directly with the dominant player in this field: NVIDIA GPUs. Comparison: AMD vs. NVIDIA GPUs for LLM Training Feature NVIDIA GPUs AMD GPUs Dominant Architecture/Platform CUDA (Compute Unified Device Architecture) – Read more