Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
Integration:  git github
9fdb69d887
Fix volumes, add notify discord, add query myself
1 year ago
26882a9f42
Turn monit on/off in Python
1 year ago
f079d590db
Attach GPU
1 year ago
471c056688
Disbale init_output_dir_from_hub_for_lora for now
11 months ago
0384d2381c
Add experiments with servereless-runpod-ggml & minotaur-15B-GGML
1 year ago
31bb9dca0e
Terminate the pod when training goes wrong and notify discord
11 months ago
64837949ba
Switch dataset to orca
11 months ago
18b1b0fa2c
Commit ft config from discord as is
1 year ago
7ce9d965ab
Initial commit
1 year ago
0c356f72d4
Minor fixes suggested by cursor
1 year ago
f564b5090a
Finish May daily picks
1 year ago
58f056bd4e
Update datasets.txt
1 year ago
31aea2a07d
Add some more models and datasets
1 year ago
Storage Buckets

README.md

You have to be logged in to leave a comment. Sign In

LLM Playground

This is a new era, with so many low-hanging fruits on the new magical tree, and so many new tools to reach them emerging every day. Things are changing so fast to play catch-up. I need a place to fix the settings and results of my experiments, this is it.

Focus Matrix

I need to focus since I don't have unlimited time and resources.

It's a focus tensor actually: tools, models, datasets, data sources, hardware, metrics, hows.

  • Tools:
    • For inference and training, use text-generation-webui if applicable, and others otherwise.
      • Focus on 4bit LoRA training with the biggest possible foundation models
      • Also try some fully-fledged training on toy models and domain datasets
    • For creating, retrieving, and preprocessing datasets
  • Models:
    • Cover the mainstream ones with a focus on the ones that don't have legal issues for commercial use.
    • Try to cover more types of LLM models to evaluate the differences.
  • Datasets:
    • Focus on the high-quality ones that enhance coding, math, reasoning, and instruction-following capacities and have no legal issues for commercial use.
    • Natural languages include English, Chinese, and other languages I like.
    • Programming languages include Python, C++, JavaScript, Rust, Lean 4, Julia, etc.
  • Data sources:
    • The data source types that can be loaded with LlamaIndex/LangChain and injected into prompts.
    • Other types I need.
  • Hardware:
    • Edge: ~8 GB RAM
    • CPU-only: ~32 GB RAM
    • Consumer-grade GPU: RTX 3090Ti/A6000 ~48G VRAM
    • Training-grade GPU: ~A100 80G VRAM
    • Mostly use cloud services to switch between hardware easily
  • Metrics:
    • Light-weight metrics to evaluate datasets for quality and diversity
    • A small set of instructions and interactions for quick sanity-check on models and easy to run like unit tests
    • Light-weight metrics to evaluate models for general and domain-specific tasks
  • Hows:
    • Experiments and visualizations to better understand why and how LLMs work

The settings

  • Base on a bare-bone docker image like nvidia/cuda:11.8.0-devel-ubuntu22.04 so I can switch between different cloud services easily
  • These files are at the root or they could be folder-specific
    • .env for environment variables
    • packages.txt for apt packages
    • requirements.txt for Python packages
    • datasets.txt for datasets
    • models.txt for models
  • The Jupyter notebooks should
    • Clone this repo and use helper scripts to fix other dependencies like text-generation-webui, GPTQ-for-LLaMa, alpaca_lora_4bit, etc.
    • Use helper scripts to download datasets and models
    • Upload results and models to HuggingFace for persistence
    • Rely on credentials for Github, HuggingFace, etc.
Tip!

Press p or to see the previous file or, n or to see the next file

About

No description

Collaborators 1

Comments

Loading...