AI & ML interests
Asynchronous Telemedicine AI Pipelines, Ambient Emotion IoT AI, Open Source for Smart Communities, AGI and ML Pipelines, Behavior Cognitive and Memory AI, Clinical Medical and Nursing AI, Genomics AI, GAN Gaming GAIL AR VR XR and Simulation AI, Graph Ontology KR KE AI, Languages and NLP AI, Quantum Compute GPU TPU NPU AI, Vision Image Document and Audio/Video AI
Classroom Examples for Today: πExamples
π Two easy ways to turbo boost your AI learning journey! π»
π AI Pair Programming
Open 2 Browsers to:
- π ChatGPT URL or URL2 and
- π Huggingface URL in separate browser windows.
- π€ Use prompts to generate a streamlit program on Huggingface or locally to test it.
- π§ For advanced work, add Python 3.10 and VSCode locally, and debug as gradio or streamlit apps.
- π Use these two superpower processes to reduce the time it takes you to make a new AI program! β±οΈ
Example Starter Prompt:
Write a streamlit program that demonstrates Data synthesis. Synthesize data from multiple sources to create new datasets. Use two datasets and demonstrate pandas dataframe query merge and join with two datasets in python list dictionaries: List of Hospitals that are over 1000 bed count by city and state, and State population size and square miles. Perform a calculated function on the merged dataset.
π₯ YouTube University Method:
- ποΈββοΈ Plan two hours each weekday to exercise your body and brain.
- π¬ Make a playlist of videos you want to learn from on YouTube. Save the links to edit later.
- π Try watching the videos at a faster speed while exercising, and sample the first five minutes of each video.
- π Reorder the playlist so the most useful videos are at the front, and take breaks to exercise.
- π Practice note-taking in markdown to instantly save what you want to remember. Share your notes with others!
- π₯ AI Pair Programming Using Long Answer Language Models with Human Feedback:
π₯ 2023 AI/ML Advanced Learning Playlists:
- 2023 QA Models and Long Form Question Answering NLP
- FHIR Bioinformatics Development Using AI/ML and Python, Streamlit, and Gradio - 2022
- 2023 ChatGPT for Coding Assistant Streamlit, Gradio and Python Apps
- 2023 BigScience Bloom - Large Language Model for AI Systems and NLP
- 2023 Streamlit Pro Tips for AI UI UX for Data Science, Engineering, and Mathematics
- 2023 Fun, New and Interesting AI, Videos, and AI/ML Techniques
- 2023 Best Minds in AGI AI Gamification and Large Language Models
- 2023 State of the Art for Vision Image Classification, Text Classification and Regression, Extractive Question Answering and Tabular Classification
- 2023 AutoML DataRobot and AI Platforms for Building Models, Features, Test, and Transparency
Language Models π£οΈ
π Bloom sets new record for most performant and efficient AI model in science! πΈ
Comparison of Large Language Models
Model Name | Model Size (in Parameters) |
---|---|
BigScience-tr11-176B | 176 billion |
GPT-3 | 175 billion |
OpenAI's DALL-E 2.0 | 500 million |
NVIDIA's Megatron | 8.3 billion |
Transformer-XL | 250 million |
XLNet | 210 million |
ChatGPT Datasets π
- WebText
- Common Crawl
- BooksCorpus
- English Wikipedia
- Toronto Books Corpus
- OpenWebText
ChatGPT Datasets - Details π
- WebText: A dataset of web pages crawled from domains on the Alexa top 5,000 list. This dataset was used to pretrain GPT-2.
- Common Crawl: A dataset of web pages from a variety of domains, which is updated regularly. This dataset was used to pretrain GPT-3.
- Language Models are Few-Shot Learners by Brown et al.
- BooksCorpus: A dataset of over 11,000 books from a variety of genres.
- Scalable Methods for 8 Billion Token Language Modeling by Zhu et al.
- English Wikipedia: A dump of the English-language Wikipedia as of 2018, with articles from 2001-2017.
- Improving Language Understanding by Generative Pre-Training Space for Wikipedia Search
- Toronto Books Corpus: A dataset of over 7,000 books from a variety of genres, collected by the University of Toronto.
- OpenWebText: A dataset of web pages that were filtered to remove content that was likely to be low-quality or spammy. This dataset was used to pretrain GPT-3.
- Language Models are Few-Shot Learners by Brown et al.
Big Science Model π
π Papers:
- BLOOM: A 176B-Parameter Open-Access Multilingual Language Model Paper
- Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism Paper
- 8-bit Optimizers via Block-wise Quantization Paper
- Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation Paper
- Other papers related to Big Science
- 217 other models optimized for use with Bloom
π Datasets:
Datasets:
- Universal Dependencies: A collection of annotated corpora for natural language processing in a range of languages, with a focus on dependency parsing.
- WMT 2014: The fourth edition of the Workshop on Statistical Machine Translation, featuring shared tasks on translating between English and various other languages.
- The Pile: An English language corpus of diverse text, sourced from various places on the internet.
- HumanEval: A dataset of English sentences, annotated with human judgments on a range of linguistic qualities.
- HumanEval: An Evaluation Benchmark for Language Understanding by Gabriel Ilharco, Daniel Loureiro, Pedro Rodriguez, and Afonso Mendes.
- FLORES-101: A dataset of parallel sentences in 101 languages, designed for multilingual machine translation.
- FLORES-101: A Massively Multilingual Parallel Corpus for Language Understanding by Aman Madaan, Shruti Rijhwani, Raghav Gupta, and Mitesh M. Khapra.
- CrowS-Pairs: A dataset of sentence pairs, designed for evaluating the plausibility of generated text.
- CrowS-Pairs: A Challenge Dataset for Plausible Plausibility Judgments by Andrea Madotto, Zhaojiang Lin, Chien-Sheng Wu, Pascale Fung, and Caiming Xiong.
- WikiLingua: A dataset of parallel sentences in 75 languages, sourced from Wikipedia.
- WikiLingua: A New Benchmark Dataset for Cross-Lingual Wikification by Jiarui Yao, Yanqiao Zhu, Ruihan Bao, Guosheng Lin, Lidong Bing, and Bei Shi.
- MTEB: A dataset of English sentences, annotated with their entailment relationships with respect to other sentences.
- Multi-Task Evaluation Benchmark for Natural Language Inference by MichaΕ Lukasik, Marcin Junczys-Dowmunt, and Houda Bouamor.
- xP3: A dataset of English sentences, annotated with their paraphrase relationships with respect to other sentences.
- xP3: A Large-Scale Evaluation Benchmark for Paraphrase Identification in Context by Aniket Didolkar, James Mayfield, Markus Saers, and Jason Baldridge.
- DiaBLa: A dataset of English dialogue, annotated with dialogue acts.
A Large-Scale Corpus for Conversation Disentanglement by Samuel Broscheit, AntΓ³nio Branco, and AndrΓ© F. T. Martins.
π Dataset Papers with Code
Deep RL ML Strategy π§
The AI strategies are:
- Language Model Preparation using Human Augmented with Supervised Fine Tuning π€
- Reward Model Training with Prompts Dataset Multi-Model Generate Data to Rank π
- Fine Tuning with Reinforcement Reward and Distance Distribution Regret Score π―
- Proximal Policy Optimization Fine Tuning π€
- Variations - Preference Model Pretraining π€
- Use Ranking Datasets Sentiment - Thumbs Up/Down, Distribution π
- Online Version Getting Feedback π¬
- OpenAI - InstructGPT - Humans generate LM Training Text π
- DeepMind - Advantage Actor Critic Sparrow, GopherCite π¦
- Reward Model Human Prefence Feedback π
For more information on specific techniques and implementations, check out the following resources:
- OpenAI's paper on GPT-3 which details their Language Model Preparation approach
- DeepMind's paper on SAC which describes the Advantage Actor Critic algorithm
- OpenAI's paper on Reward Learning which explains their approach to training Reward Models
- OpenAI's blog post on GPT-3's fine-tuning process