a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
-
Updated
Jun 25, 2024 - Jupyter Notebook
a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
This repository features a custom-built decoder-only language model (LLM) with a total of 37 million parameters 🔥. I train the model to be able to ask question from a given context
Experimental project for AI and NLP based on Transformer Architecture
Generate caption on images using CNN Encoder- LSTM Decoder structure
Transformers Intuition
Implementation of the GPT-3 paper: Language Models are Few-Shot Learners
An explainable and simplified version of OLMo model
a dna sequence generation/classification using transformers
An LLM based tool for generation of cheese advirtisements
This project aims to simplify texts from research papers using advanced natural language processing (NLP) techniques, making them more accessible to a broader audience
Using LLMs in huggingface for sentiment analysis, translation, summarization and extractive question answering
A multimodal vision model that takes in an image and a prompt query, and output the answer
Mimic GPT-2 (124 million parameter) like Transformer (decoder) model. Compare performance against OpenAI's Model for prediction accuracy.
A miniGPT inspired from the original NanoGPT released by OpenAI. This is a notebook to walk through the decoder part of the transformer architecture with details outlined.
On the Design and Performance of Machine Learning Based Error Correcting Decoders
A mini version of GPT implemented on shakespear using BPE
Decoder model for language modelling
Build Text summarizer for arabic language
Text Generation using RNN, LSTM, and Transformer
Add a description, image, and links to the decoder-model topic page so that developers can more easily learn about it.
To associate your repository with the decoder-model topic, visit your repo's landing page and select "manage topics."