Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fixed all "Open in Colab" absolute paths #850

Merged
merged 5 commits into from
Jan 16, 2025
Merged
Show file tree
Hide file tree
Changes from 4 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -8,8 +8,7 @@
"source": [
"# Using Amazon Bedrock with Llama\n",
"\n",
"Open this notebook in <a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/llama_api_providers/examples_with_aws/getting_started_llama2_on_amazon_bedrock.ipynb\"><img data-canonical-src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\" src=\"https://camo.githubusercontent.com/f5e0d0538a9c2972b5d413e0ace04cecd8efd828d133133933dfffec282a4e1b/68747470733a2f2f636f6c61622e72657365617263682e676f6f676c652e636f6d2f6173736574732f636f6c61622d62616467652e737667\"></a>\n",
"\n",
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/aws/getting_started_llama_3_on_amazon_bedrock.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"\n",
"Use this notebook to quickly get started with Llama on Bedrock. You can access the Amazon Bedrock API using the AWS Python SDK.\n",
"\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,7 @@
"source": [
"# Prompt Engineering with Llama 2 - Using Amazon Bedrock + LangChain\n",
"\n",
"Open this notebook in <a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/quickstart/Prompt_Engineering_with_Llama_2.ipynb\"><img data-canonical-src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\" src=\"https://camo.githubusercontent.com/f5e0d0538a9c2972b5d413e0ace04cecd8efd828d133133933dfffec282a4e1b/68747470733a2f2f636f6c61622e72657365617263682e676f6f676c652e636f6d2f6173736574732f636f6c61622d62616467652e737667\"></a>\n",
"\n",
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/aws/prompt_engineering_with_llama_2_on_amazon_bedrock.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"\n",
"Prompt engineering is using natural language to produce a desired response from a large language model (LLM).\n",
"\n",
Expand Down
2 changes: 1 addition & 1 deletion 3p-integrations/aws/react_llama_3_bedrock_wk.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@
"# Advanced Techniques\n",
"## 1. ReAct\n",
"\n",
"Open this notebook in <a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/llama_api_providers/examples_with_aws/ReAct_Llama_2_Bedrock-WK.ipynb\"><img data-canonical-src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\" src=\"https://camo.githubusercontent.com/f5e0d0538a9c2972b5d413e0ace04cecd8efd828d133133933dfffec282a4e1b/68747470733a2f2f636f6c61622e72657365617263682e676f6f676c652e636f6d2f6173736574732f636f6c61622d62616467652e737667\"></a>\n",
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/aws/react_llama_3_bedrock_wk.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"\n",
"LLMs abilities for reasoning (e.g. chain-of-thought CoT prompting) and acting have primarily been studied as separate topics. **ReAct** [Shunyu Yao et al. ICLR 2023](https://arxiv.org/pdf/2210.03629.pdf) (Reason and Act) is a method to generate both reasoning traces and task-specific actions in an interleaved manner.\n",
"\n",
Expand Down
4 changes: 1 addition & 3 deletions 3p-integrations/e2b-ai-analyst/README.md
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
# AI Analyst with Llama and E2B
This is an AI-powered code and data analysis tool powered by Meta Llama and the [E2B SDK](https://e2b.dev/docs).

![Preview](https://private-user-images.githubusercontent.com/33395784/382129362-3bc8b017-4a09-416c-b55c-ce53da7e5560.gif?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MzEwNzg5NDQsIm5iZiI6MTczMTA3ODY0NCwicGF0aCI6Ii8zMzM5NTc4NC8zODIxMjkzNjItM2JjOGIwMTctNGEwOS00MTZjLWI1NWMtY2U1M2RhN2U1NTYwLmdpZj9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDExMDglMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQxMTA4VDE1MTA0NFomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWRlZjc4ZDhmNzY1ZjExODM4OGEzMzFiZGRjMWU4ZmQ0OWI5MjJmNTY5MGZkZDk4MmI1MWRiMzFhY2UxYjM0NjcmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.Kaab1leGbgBiQhfh6bV1VHTc8QfinP7ufqwmn7Ra57c)

→ Try on [ai-analyst.e2b.dev](https://ai-analyst.e2b.dev/)

## Features
Expand Down Expand Up @@ -54,4 +52,4 @@ In `.env.local`, add an API key for at least one LLM provider:

For Ollama, provide the base URL instead of the API key:

- Ollama: `OLLAMA_BASE_URL`
- Ollama: `OLLAMA_BASE_URL`
3 changes: 2 additions & 1 deletion 3p-integrations/groq/llama3_cookbook_groq.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,8 @@
"source": [
"# Llama 3 Cookbook with LlamaIndex and Groq\n",
"\n",
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/llama_api_providers/llama3_cookbook_groq.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/groq/llama3_cookbook_groq.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"\n",
"\n",
"Meta developed and released the Meta [Llama 3](https://ai.meta.com/blog/meta-llama-3/) family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks.\n",
"\n",
Expand Down
2 changes: 1 addition & 1 deletion 3p-integrations/langchain/langgraph_rag_agent.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"id": "6912ab05-f66a-40a9-a4a5-4deb80d2e0d9",
"metadata": {},
"source": [
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/3p_integrations/langchain/langgraph_rag_agent.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/langchain/langgraph_rag_agent.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion 3p-integrations/langchain/langgraph_rag_agent_local.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"id": "1f53f753-12c6-4fac-b910-6e96677d8a49",
"metadata": {},
"source": [
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/3p_integrations/langchain/langgraph_rag_agent_local.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/langchain/langgraph_rag_agent_local.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
]
},
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"id": "8ac4ba3b-c438-4f2e-8f52-39846beb5642",
"metadata": {},
"source": [
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/3p_integrations/langchain/langgraph_tool_calling_agent.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/langchain/langgraph_tool_calling_agent.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
]
},
{
Expand Down
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Screenshot 2024-05-30 at 10.53.54 AM.png
is broken in colab, note to self

Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/3p_integrations/llamaindex/dlai_agentic_rag/Building_Agentic_RAG_with_Llamaindex_L2_Tool_Calling.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/llamaindex/dlai_agentic_rag/Building_Agentic_RAG_with_Llamaindex_L2_Tool_Calling.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"\n",
"This notebook ports the DeepLearning.AI short course [Building Agentic RAG with Llamaindex Lesson 2 Tool Calling](https://learn.deeplearning.ai/courses/building-agentic-rag-with-llamaindex/lesson/3/tool-calling) to using Llama 3. It shows how to use Llama 3 to not only pick a function to execute, but also infer an argument to pass through the function.\n",
"\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/3p_integrations/llamaindex/dlai_agentic_rag/Building_Agentic_RAG_with_Llamaindex_L3_Building_an_Agent_Reasoning_Loop.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/llamaindex/dlai_agentic_rag/Building_Agentic_RAG_with_Llamaindex_L3_Building_an_Agent_Reasoning_Loop.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"\n",
"This notebook ports the DeepLearning.AI short course [Building Agentic RAG with Llamaindex Lesson 3 Building an Agent Reasoning Loop](https://learn.deeplearning.ai/courses/building-agentic-rag-with-llamaindex/lesson/4/building-an-agent-reasoning-loop) to using Llama 3. It shows how to define a complete agent reasoning loop to reason over tools and multiple steps on a complex question the user asks about a single document while maintaining memory.\n",
"\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/3p_integrations/llamaindex/dlai_agentic_rag/Building_Agentic_RAG_with_Llamaindex_L4_Building_a_Multi-Document_Agent.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"<a href=\"https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/llamaindex/dlai_agentic_rag/Building_Agentic_RAG_with_Llamaindex_L4_Building_a_Multi-Document_Agent.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>\n",
"\n",
"This notebook ports the DeepLearning.AI short course [Building Agentic RAG with Llamaindex Lesson 4 Building a Multi-Document Agent](https://learn.deeplearning.ai/courses/building-agentic-rag-with-llamaindex/lesson/5/building-a-multi-document-agent) to using Llama 3. It shows how to use an agent to handle multiple documents and increasing degrees of complexity.\n",
"\n",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"metadata": {},
"source": [
"# Generating Knowledge Graphs with LLMs and Structured Outputs\n",
"[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/3p_integrations/togetherai/knowledge_graphs_with_structured_outputs.ipynb)"
"[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/togetherai/knowledge_graphs_with_structured_outputs.ipynb)"
]
},
{
Expand All @@ -30,60 +30,15 @@
},
{
"cell_type": "code",
"execution_count": 1,
"execution_count": null,
"metadata": {
"colab": {
"base_uri": "https://localhost:8080/"
},
"id": "DFAjay1FZVrn",
"outputId": "d4b17b31-c125-4de5-ad54-6d4d08d81eaa"
},
"outputs": [
{
"name": "stdout",
"output_type": "stream",
"text": [
"Requirement already satisfied: together in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (1.3.3)\n",
"Requirement already satisfied: aiohttp<4.0.0,>=3.9.3 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (3.10.10)\n",
"Requirement already satisfied: click<9.0.0,>=8.1.7 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (8.1.7)\n",
"Requirement already satisfied: eval-type-backport<0.3.0,>=0.1.3 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (0.2.0)\n",
"Requirement already satisfied: filelock<4.0.0,>=3.13.1 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (3.16.1)\n",
"Requirement already satisfied: numpy>=1.23.5 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (1.26.4)\n",
"Requirement already satisfied: pillow<11.0.0,>=10.3.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (10.4.0)\n",
"Requirement already satisfied: pyarrow>=10.0.1 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (18.0.0)\n",
"Requirement already satisfied: pydantic<3.0.0,>=2.6.3 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (2.9.2)\n",
"Requirement already satisfied: requests<3.0.0,>=2.31.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (2.32.3)\n",
"Requirement already satisfied: rich<14.0.0,>=13.8.1 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (13.9.3)\n",
"Requirement already satisfied: tabulate<0.10.0,>=0.9.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (0.9.0)\n",
"Requirement already satisfied: tqdm<5.0.0,>=4.66.2 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (4.66.6)\n",
"Requirement already satisfied: typer<0.13,>=0.9 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from together) (0.12.5)\n",
"Requirement already satisfied: aiohappyeyeballs>=2.3.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from aiohttp<4.0.0,>=3.9.3->together) (2.4.3)\n",
"Requirement already satisfied: aiosignal>=1.1.2 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from aiohttp<4.0.0,>=3.9.3->together) (1.3.1)\n",
"Requirement already satisfied: attrs>=17.3.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from aiohttp<4.0.0,>=3.9.3->together) (24.2.0)\n",
"Requirement already satisfied: frozenlist>=1.1.1 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from aiohttp<4.0.0,>=3.9.3->together) (1.5.0)\n",
"Requirement already satisfied: multidict<7.0,>=4.5 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from aiohttp<4.0.0,>=3.9.3->together) (6.1.0)\n",
"Requirement already satisfied: yarl<2.0,>=1.12.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from aiohttp<4.0.0,>=3.9.3->together) (1.17.1)\n",
"Requirement already satisfied: async-timeout<5.0,>=4.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from aiohttp<4.0.0,>=3.9.3->together) (4.0.3)\n",
"Requirement already satisfied: annotated-types>=0.6.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from pydantic<3.0.0,>=2.6.3->together) (0.7.0)\n",
"Requirement already satisfied: pydantic-core==2.23.4 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from pydantic<3.0.0,>=2.6.3->together) (2.23.4)\n",
"Requirement already satisfied: typing-extensions>=4.6.1 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from pydantic<3.0.0,>=2.6.3->together) (4.12.2)\n",
"Requirement already satisfied: charset-normalizer<4,>=2 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from requests<3.0.0,>=2.31.0->together) (3.4.0)\n",
"Requirement already satisfied: idna<4,>=2.5 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from requests<3.0.0,>=2.31.0->together) (3.10)\n",
"Requirement already satisfied: urllib3<3,>=1.21.1 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from requests<3.0.0,>=2.31.0->together) (2.2.3)\n",
"Requirement already satisfied: certifi>=2017.4.17 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from requests<3.0.0,>=2.31.0->together) (2024.8.30)\n",
"Requirement already satisfied: markdown-it-py>=2.2.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from rich<14.0.0,>=13.8.1->together) (3.0.0)\n",
"Requirement already satisfied: pygments<3.0.0,>=2.13.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from rich<14.0.0,>=13.8.1->together) (2.18.0)\n",
"Requirement already satisfied: shellingham>=1.3.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from typer<0.13,>=0.9->together) (1.5.4)\n",
"Requirement already satisfied: mdurl~=0.1 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from markdown-it-py>=2.2.0->rich<14.0.0,>=13.8.1->together) (0.1.2)\n",
"Requirement already satisfied: propcache>=0.2.0 in /Users/jeffxtang/anaconda3/envs/llama-recipes/lib/python3.10/site-packages (from yarl<2.0,>=1.12.0->aiohttp<4.0.0,>=3.9.3->together) (0.2.0)\n",
"Collecting graphviz\n",
" Using cached graphviz-0.20.3-py3-none-any.whl.metadata (12 kB)\n",
"Using cached graphviz-0.20.3-py3-none-any.whl (47 kB)\n",
"Installing collected packages: graphviz\n",
"Successfully installed graphviz-0.20.3\n"
]
}
],
"outputs": [],
"source": [
"!pip install together\n",
"!pip install graphviz"
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/3p_integrations/togetherai/pdf_to_podcast_using_llama_on_together.ipynb)"
"[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/meta-llama/llama-cookbook/blob/main/3p-integrations/togetherai/pdf_to_podcast_using_llama_on_together.ipynb)"
]
},
{
Expand Down
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ This repository covers the most popular community approaches, use-cases and the
> * [Text to SQL](./end-to-end-use-cases/coding/text2sql/)


> Note: We recently did a refactor of the repo, [archive-main](https://github.com/meta-llama/llama-recipes/tree/archive-main) is a snapshot branch from before the refactor
> Note: We recently did a refactor of the repo, [archive-main](https://github.com/meta-llama/llama-cookbook/tree/archive-main) is a snapshot branch from before the refactor

## Repository Structure:

Expand All @@ -44,7 +44,7 @@ A: Checkout the Fine-Tuning FAQ [here](./src/docs/)

- Q: Some links are broken/folders are missing:

A: We recently did a refactor of the repo, [archive-main](https://github.com/meta-llama/llama-recipes/tree/archive-main) is a snapshot branch from before the refactor
A: We recently did a refactor of the repo, [archive-main](https://github.com/meta-llama/llama-cookbook/tree/archive-main) is a snapshot branch from before the refactor

- Where can we find details about the latest models?

Expand Down
Loading