Skip to content

Commit 4ef43b8

Browse files
committed
fix links
1 parent 3fe85db commit 4ef43b8

File tree

6 files changed

+9
-9
lines changed

6 files changed

+9
-9
lines changed

_config.yml

+1
Original file line numberDiff line numberDiff line change
@@ -139,6 +139,7 @@ sphinx:
139139
- https://www.crcv.ucf.edu/data/UCF101.php
140140
- https://www.pinecone.io/learn/series/faiss/locality-sensitive-hashing
141141
- https://chat.lmsys.org
142+
- https://platform.openai.com/docs/.*
142143
linkcheck_allowed_redirects:
143144
https://doi.org/.*/.*: https://.*
144145
https://codespaces.new/.*: https://github.com/login.*

index.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -225,7 +225,7 @@ LLM
225225
A [Large Language Model](https://en.wikipedia.org/wiki/Large_language_model) is neural network (often a {term}`transformer` containing billions of parameters) designed to perform tasks in natural language via [fine tuning](<https://en.wikipedia.org/wiki/Fine-tuning_(machine_learning)>) or [prompt engineering](https://en.wikipedia.org/wiki/Prompt_engineering).
226226
227227
MLOps
228-
[Machine Learning Operations](https://blogs.nvidia.com/blog/2020/09/03/what-is-mlops): best practices to run AI using software products & cloud services
228+
[Machine Learning Operations](https://blogs.nvidia.com/blog/what-is-mlops): best practices to run AI using software products & cloud services
229229
230230
MoE
231231
[Mixture-of-Experts](https://en.wikipedia.org/wiki/Mixture_of_experts) is a technique which uses one or more specialist model(s) from a collection of models ("experts") to solve general problems. Not that this is different from [ensemble](https://en.wikipedia.org/wiki/Ensemble_learning) models (which combine results from all models).

mlops-engines.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ Cons:
8484

8585
## vLLM
8686

87-
This is an open-source project created by researchers at Berkeley to improve the performance of LLM inferencing. [vLLM](https://vllm.ai) primarily optimises LLM throughput via methods like PagedAttention and Continuous Batching. The project is fairly new and there is ongoing development.
87+
This is an open-source project created by researchers at Berkeley to improve the performance of LLM inferencing. https://github.com/vllm-project/vllm primarily optimises LLM throughput via methods like PagedAttention and Continuous Batching. The project is fairly new and there is ongoing development.
8888

8989
Pros:
9090

models.md

+5-5
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@
77
Some ideas:
88
99
- [The History of Open-Source LLMs: Better Base Models (part 2)](https://cameronrwolfe.substack.com/p/the-history-of-open-source-llms-better) (LLaMA, MPT, Falcon, LLaMA-2)
10-
- [Papers I've read this week, Mixture of Experts edition](https://finbarrtimbers.substack.com/p/papers-ive-read-this-week-mixture) (conditional routing models)
10+
- [Papers I've read this week, Mixture of Experts edition](https://www.artfintel.com/p/papers-ive-read-this-week-mixture) (conditional routing models)
1111
- [AI and Memory Wall](https://medium.com/riselab/ai-and-memory-wall-2cb4265cb0b8)
1212
- https://github.com/imaurer/awesome-decentralized-llm
1313
- https://github.com/huggingface/transformers/blob/main/awesome-transformers.md
@@ -45,7 +45,7 @@ ChatGPT is sensitive to tweaks to the input phrasing or attempting the same prom
4545

4646
#### GPT-4
4747

48-
[GPT-4 is a language model developed by OpenAI](https://openai.com/research/gpt-4). It is the successor to GPT-3 and has been made publicly available via the paid chatbot product ChatGPT Plus and via OpenAI's API. It is a large multimodal model that can accept image and text inputs and emit text outputs, [though multimodal capabilities aren't released to the public yet](https://analyticsindiamag.com/what-happened-to-multimodal-gpt-4). It exhibits human-level performance on various professional and academic benchmarks and can follow complex instructions in natural language and solve difficult problems with accuracy. It can handle input prompts of up to 32k tokens, which is a significant increase from GPT-3.5's 4k tokens. It can solve complex mathematical and scientific problems beyond the capabilities of GPT-3.5, such as advanced calculus problems or simulating chemical reactions [more effectively than its predecessor](https://www.searchenginejournal.com/gpt-4-vs-gpt-3-5/482463). It is more reliable, creative, and able to handle much more nuanced instructions than GPT-3.5.
48+
[GPT-4 is a language model developed by OpenAI](https://openai.com/research/gpt-4). It is the successor to GPT-3 and has been made publicly available via the paid chatbot product ChatGPT Plus and via OpenAI's API. It is a large multimodal model that can accept image and text inputs and emit text outputs, [though multimodal capabilities aren't released to the public yet](http://analyticsindiamag.com/what-happened-to-multimodal-gpt-4/). It exhibits human-level performance on various professional and academic benchmarks and can follow complex instructions in natural language and solve difficult problems with accuracy. It can handle input prompts of up to 32k tokens, which is a significant increase from GPT-3.5's 4k tokens. It can solve complex mathematical and scientific problems beyond the capabilities of GPT-3.5, such as advanced calculus problems or simulating chemical reactions [more effectively than its predecessor](https://www.searchenginejournal.com/gpt-4-vs-gpt-3-5/482463). It is more reliable, creative, and able to handle much more nuanced instructions than GPT-3.5.
4949

5050
Despite its capabilities, [GPT-4 still sometimes "hallucinates"](https://www.reddit.com/r/ChatGPT/comments/12fmrcd/examples_of_gpt4_hallucination) facts and makes reasoning errors.
5151

@@ -111,7 +111,7 @@ Before [](#chatgpt)'s (GPT-3.5) public release we had [GPT-3](https://en.wikiped
111111
There has been a few visible marks across modalities of AI models, highly catalysing growth of open source:
112112

113113
- [Meta AI launches LLaMA](https://ai.meta.com/blog/large-language-model-llama-meta-ai), open sourcing the code but not the weights.
114-
- [StabilityAI released Stable Diffusion](https://stability.ai/blog/stable-diffusion-announcement).
114+
- [StabilityAI released Stable Diffusion](https://stability.ai/news/stable-diffusion-announcement).
115115

116116
#### [Stable Diffusion](https://registry.premai.io/detail.html?service=stable-diffusion-1-5)
117117

@@ -371,15 +371,15 @@ Till now we've mostly been looking at LLMs in general and not other models, let'
371371
372372
#### [Stable Diffusion XL](https://registry.premai.io/detail.html?service=stable-diffusion-xl-with-refiner)
373373
374-
[StabilityAI released Stable Diffusion XL 1.0 (SDXL)](https://stability.ai/blog/stable-diffusion-sdxl-1-announcement) models on 26th July, being current State of the Art for text-to-image and image-to-image generation open sourced models. They released a [base model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0) and a [refinement model](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0) which is used to improve the visual fidelity of samples generated by SDXL.
374+
[StabilityAI released Stable Diffusion XL 1.0 (SDXL)](https://stability.ai/news/stable-diffusion-sdxl-1-announcement) models on 26th July, being current State of the Art for text-to-image and image-to-image generation open sourced models. They released a [base model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0) and a [refinement model](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0) which is used to improve the visual fidelity of samples generated by SDXL.
375375
376376
Few months back they released Stable-diffusion-xl {cite}`podell2023sdxl` [base](https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9) and [refinement](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-0.9) models versioned as 0.9, where license permitting only research purpose usages.
377377
378378
SDXL consistently surpasses all previous versions of Stable Diffusion models by a significant margin:
379379
380380
```{figure} https://static.premai.io/book/models_sdxl-winrate.png
381381
:width: 60%
382-
[SDXL Winrate](https://stability.ai/blog/stable-diffusion-sdxl-1-announcement)
382+
[SDXL Winrate](https://stability.ai/news/stable-diffusion-sdxl-1-announcement)
383383
```
384384

385385
##### Uniqueness

references.md

-1
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,6 @@
77
88
- "Catching up on the weird world of LLMs" (summary of the last few years) https://simonwillison.net/2023/Aug/3/weird-world-of-llms
99
- "Open challenges in LLM research" (exciting post title but mediocre content) https://huyenchip.com/2023/08/16/llm-research-open-challenges.html
10-
- "AI forest" https://www.michaeldempsey.me/blog/2023/07/18/the-dark-forest-of-rd-and-capital-deployment-in-ai
1110
- https://github.com/zeno-ml/zeno-build/tree/main/examples/analysis_gpt_mt/report
1211
- "Patterns for Building LLM-based Systems & Products" (Evals, RAG, fine-tuning, caching, guardrails, defensive UX, and collecting user feedback) https://eugeneyan.com/writing/llm-patterns
1312

sdk.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@ The list of vector stores that LangChain supports can be found [here](https://ap
4646

4747
### Models
4848

49-
This is the heart of most LLMs, where the core functionality resides. There are broadly [2 different types of models](https://python.langchain.com/docs/modules/model_io/models) which LangChain integrates with:
49+
This is the heart of most LLMs, where the core functionality resides. There are broadly [2 different types of models](https://python.langchain.com/docs/modules/model_io) which LangChain integrates with:
5050

5151
- **Language**: Inputs & outputs are `string`s
5252
- **Chat**: Run on top of a Language model. Inputs are a list of chat messages, and output is a chat message

0 commit comments

Comments
 (0)