power-greg commited on
Commit
ce40565
1 Parent(s): a97f966

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -1
README.md CHANGED
@@ -7,4 +7,34 @@ sdk: static
7
  pinned: false
8
  ---
9
 
10
- Edit this `README.md` markdown file to author your organization card 🔥
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  pinned: false
8
  ---
9
 
10
+ # Introducing Lamini, the LLM Engine for Rapid Customization
11
+
12
+ [Lamini](lamini.ai) gives every developer the superpowers that took the world from GPT-3 to ChatGPT!
13
+
14
+ Today, you can try out our open dataset generator for training instruction-following LLMs (like ChatGPT) on [Github](https://lamini.ai/).
15
+
16
+ [Sign up](https://lamini.ai/contact) for early access to our full LLM training module, including enterprise features like cloud prem deployments.
17
+
18
+
19
+ # Training LLMs should be as easy as prompt-tuning 🦾
20
+ Why is writing a prompt so easy, but training an LLM from a base model still so hard? Iteration cycles for finetuning on modest datasets are measured in months because it takes significant time to figure out why finetuned models fail. Conversely, prompt-tuning iterations are on the order of seconds, but performance plateaus in a matter of hours. Only a limited amount of data can be crammed into the prompt, not the terabytes of data in a warehouse.
21
+
22
+ It took OpenAI months with an incredible ML team to fine-tune and run RLHF on their base GPT-3 model that was available for years — creating what became ChatGPT. This training process is only accessible to large ML teams, often with PhDs in AI.
23
+
24
+ Technical leaders at Fortune 500 companies have told us:
25
+
26
+ * “Our team of 10 machine learning engineers hit the OpenAI finetuning API, but our model got worse — help!”
27
+ * “I don’t know how to make the best use of my data — I’ve exhausted all the prompt magic we can summon from tutorials online.”
28
+ That’s why we’re building Lamini: to give every developer the superpowers that took the world from GPT-3 to ChatGPT.
29
+
30
+
31
+ # Rapidly train LLMs to be as good as ChatGPT from any base model 🚀
32
+ Lamini is an LLM engine that allows any developer, not just machine learning experts, to train high-performing LLMs on large datasets using the Lamini library.
33
+
34
+ The optimizations in this library reach far beyond what’s available to developers now, from more challenging ones like RLHF to simpler ones like reducing hallucinations.
35
+
36
+ ![Lamini Process Step by Step](lamini.png "Lamini LLM Engine")
37
+
38
+ Lamini runs across platforms, from OpenAI’s models to open-source ones on HuggingFace, with more to come soon. We are agnostic to base models, as long as there’s a way for our engine to train and run them. In fact, Lamini makes it easy to run multiple base model comparisons in just a single line of code.
39
+
40
+ Now that you know a bit about where we’re going, today, we’re excited to release our first major community resource!