current state of llms
intro
- clarify the "real" state of llms in terms of capabilities and real-world uses
- clarify my own personal viewpoints on llms
- clarify my own personal viewpoints on llm-creating companies themselves
what is an llm
text = "the house is red"
tokens = tokenize(text)
tokens is now [54, 256, 12, 28]
result = []
while (next_token := get_next_token(tokens)) != STOP_TOKEN:
result += next_token
tokens += result
result is now [278, 279, 300, 301, 302, 304]
result_text = decode(result)
result_text is now "Which house are you referring to?"
obviously, this code doesn't work, its pseudocode lol. but it gets my point across:
- our input text is transformed into a list of tokens.
- the llm is doing fancy math to figure out the next token.
- these tokens are generated till a special stop token is emitted.
- the result is decoded to regular text.
how capable are llms?
capable
not capable
enhancing llm capabilities
rag - retrieval augmentation generation
- retrieve the answer to a user's input from a pre-made database
- give this answer to an llm + tell it to generate an answer based on the query and the answer from the database
finetuning
does it change how capable an llm is?
my views on llms
usage for learning
usage for doing (?)
time_saved = time_you_would_take_to_do_task - time_llm_takes_to_do_task
if (task_important)
{
time_saved += time_to_verify
// here, time_saved == time_you_would_take_to_do_task
}
other usages
my view on llm companies
tldr
links :)
- [1]: https://www.youtube.com/watch?v=Q0TpWitfxPk