Introduction to Large Language Models for Statisticians
15th October 2024
ChatGPT Heralds an Intellectual Revolution
25 February 2023
By Henry Kissinger, Eric Schmidt, and Daniel Huttenlocher
Generative artificial intelligence presents a philosophical and practical challenge on a scale not experienced since the start of the Enlightenment.
A new technology bids to transform the human cognitive process as it has not been shaken up since the invention of printing. The technology that printed the Gutenberg Bible in 1455 made abstract human thought communicable generally and rapidly. But new technology today reverses that process. Whereas the printing press caused a profusion of modern human thought, the new technology achieves its distillation and elaboration. In the process, it creates a gap between human knowledge and human understanding. If we are to navigate this transformation successfully, new concepts of human thought and interaction with machines will need to be developed. This is the essential challenge of the Age of Artificial Intelligence.
…
Mr. Kissinger served as secretary of state, 1973-77, and White House national security adviser, 1969-75. Mr. Schmidt was CEO of Google, 2001-11 and executive chairman of Google and its successor, Alphabet Inc., 2011-17. Mr. Huttenlocher is dean of the Schwarzman College of Computing at the Massachusetts Institute of Technology. They are authors of “The Age of AI: And Our Human Future.” The authors thank Eleanor Runde for her research.
Refresh of browser may be needed for timeline to render correctly
Data Source: https://en.wikipedia.org/wiki/List_of_chatbots (Accessed on 11/08/2024)
chat with user
multiple purposes
new creative content
Refresh of browser may be needed for timeline to render correctly
Data Source: https://en.wikipedia.org/wiki/Large_language_model#List (Accessed on 11/08/2024)
Local LLM
GPT4All LM Studio Jan llama.cpp llamafile Ollama NextChat …
chatgpt-4o-latest
gpt-4o
gpt-4o-mini
gpt-3.5-turbo
dall-e-3
text-embedding-ada-002
tts-1-hd
whisper-1
curl https://api.openai.com/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $OPENAI_API_KEY" \
-d '{
"model": "gpt-4o-mini",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant."
},
{
"role": "user",
"content": "Hello!"
}
]
}'
llama3.2:1b
llama3.2:3b
llama3.1:8b
llama3.1:70b
llama3.1:405b
gemma2:2b
gemma2:9b
gemma2:27b
llava:7b
curl http://localhost:11434/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "llama3.2:1b",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant."
},
{
"role": "user",
"content": "Hello!"
}
]
}'
llama3.1:8b
has about 8 billion parameters.dall-e-3
generates images from user text inputllava:7b
is a multi-modal model and can take image inputmathstral:7b
is designed for math reasoning and scientific discoverydeepseek-coder-v2:16b
comparable to gpt-4-turbo
in code-specific tasksmeditron:7b
adapted from llama2
for medical domaingpt-4o
, gpt-4o-mini
and dall-e-3
with payment of US$5 so farllama3.1:8b
and llava:7b
Input: All models are wrong, but some are
Updated input:
All models are wrong, but some are useful
Updated input:
All models are wrong, but some are useful.
LLM
useful
.
<|end|>
Output: useful.
Input Where there’s a will, there’s a
Token Where there ’s a will , there ’s a
Token ID 11977 1354 802 261 738 11 1354 802 261
LLM
2006
way
summary summarise summarize summarising summarizations
summary summ ar ise summ ar ize summ ar ising summ ar izations
3861 141249 277 1096 750 5066 25434
Input: Where there’s a will, there’s a
Token Probability
2006
301
35
4443
4
…
top_p
:Token Probability
2006
301
temperature
Token Probability
2006
301
35
4443
4
…
seed
ensures the same random sample given the same input (important for reproducibility!), but the same seed
may not yield the same result across different systems.Computer Science.
Data Science.
Chain-of-thought aims to make the LLM “think” before answering.
Reasoning is a core component of human intelligence and LLM can mimic “reasoning” from memorisation and pattern matching trained from large corpus of text.
Statistics.
3.
Some common components include:
🤔 Pondering
(for later)
How do LLMs complement or hinder statistical thinking?
What role should LLMs play in decision-making processes and research?
How will LLMs impact the training and development of future statisticians and data scientists?
What are the use cases of LLM for you (if any)?
emitanaka.org/workshop-LLM-2024/