GIT repositories: All
Skip-gram-with-NS.ipynb
Skip-gram with negative sampling
#text processing #machine learning modellingKeras-GAN
Keras implementations of Generative Adversarial Networks.
#machine learning modelling #neural networksequence-to-sequence translation (by Keras)
Implementation a basic character-level sequence-to-sequence model. Applied to translating short English sentences into short French sentences, character-by-character.
#text processing #kerasLSTM stateful (by Keras)
Example demonstrate how to use a stateful LSTM model, stateful vs stateless LSTM performance comparison
#text processing #machine learning modelling #neural networkpretrained word embeddings (by Keras)
This script loads pre-trained word embeddings (GloVe embeddings) into a frozen Keras Embedding layer, and uses it to train a text classification model on the 20 Newsgroup dataset
#nlp #text processing #kerasLSTM text generation (by Keras)
Example script to generate text from Nietzsche's writings.
#nlp #text processing #machine learning modellingfastText on IMDB-dataset (by Keras)
This example demonstrates the use of fasttext for text classification
#nlp #text processing #kerasbAbI mem-NN (by Keras)
End-to-End memory network on bAbI-dataset (reading comprehension Question-Answering).
#keras #machine learning modelling #neural networkLSTM-based network on bAbI-dataset (by Keras)
Recurrent neural networks for modeling Facebook’s bAbi dataset, “a mixture of 20 tasks for testing text understanding and reasoning”
#keras #machine learning modelling #neural networkunstructured-text-modelling
Text Analytics (Unsupervised Clustering) and Neural Network Modelling
#text processing #machine learning modelling #neural networkmulti-label-text-classification
Holds code for collecting data from arXiv to build a multi-label text classification dataset and a simpler classifier on top of that.
#text processing #machine learning modellinggpt2-from-scratch
How to create GPT-2, a powerful language model developed by OpenAI from scratch that can generate human-like text by predicting the next word in a sequence.
#llm #machine learning modelling #gptCompact Language Detector 2 (c++)
CLD2 probabilistically detects over 80 languages in Unicode UTF-8 text, either plain text or HTML/XML. Legacy encodings must be converted to valid UTF-8 by the caller. For mixed-language input, CLD2 returns the top three languages found and their approximate percentages of the total text bytes
#text processing #c++counsel-chat (counsel_chat.ipynb)
This repository holds the code for working with data from counselchat.com. The scarped data are from individiuals seeking assistance from licensed therapists and their associated responses.
#llm #nlp #text processinggpt-2-simple
Python package to easily retrain OpenAI's GPT-2 text-generating model on new texts
#llm #gpt #neural networkgpt-2 (openAI) original
This repository is meant to be a starting point for researchers and engineers to experiment with GPT-2. Implementation from OpenAI
#llm #gpt #neural networkMobileLLM
This repository contains the training code of MobileLLM introduced in our work: "MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases", MobileLLM-125M/350M
#llm #ai #neural networkmml-book
Mathematics For Machine Learning. [Marc Peter Deisenroth, A Aldo Faisal, and Cheng Soon Ong]. To be published by Cambridge University Press. 🏆❇️✴️✨🎯
#machine learning modellingfastText
fastText is a library for efficient learning of word representations and sentence classification. 🖥️🤖📡💻🌐⌨️🖱️🌍⭐👩💻📱
#text processing #embeddingsOpenNN
OpenNN is a software library written in C++ for advanced analytics. It implements neural networks, the most successful machine learning method.
#c++ #neural networkopenai-finetuning-example
This repository provides an example of fine-tuning OpenAI's GPT-4o-mini model for classifying customer service support tickets. Through fine-tuning, we are able to increase the classification accuracy from 69% to 94%.
#llm #nlp #text processingtransformer-tf
Attention Is All You Need Implementation (TensorFlow 2.x) 🚀 This repository contains the TensorFlow implementation of the paper. This implementation can be used to perform any sequence to sequence task with some minimal code changes.
#tensorflow #transformerstf-transformers
Imagine auto-regressive generation to be 90x faster. tf-transformers (Tensorflow Transformers) is designed to harness the full power of Tensorflow 2, designed specifically for Transformer based architecture.
#tensorflow #transformersmgpt
We introduce mGPT, a multilingual variant of GPT-3, pretrained on 61 languages from linguistically diverse 25 language families using Wikipedia and C4 Corpus.
#llm #text processing #transformersgpt-mini
Yet another minimalistic Tensorflow (re-)re-implementation of Karpathy's Pytorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer).
#transformersml-cs335
Machine-learning mini-course: numpy examples for beginner-level
#machine learning modellinggpt3-sandbox
The goal of this project is to enable users to create cool web demos using the newly released OpenAI GPT-3 API with just a few lines of Python.
gpt-3-simple-tutorial
Generate SQL from Natural Language Sentences using OpenAI's GPT-3 Model
miniGPT
Minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer), both training and inference
#llm #nlp #text processing