Meta dropped Llama-3 just a few weeks ago and it has taken everyone by surprise. People are coming up with wild use cases every day, pushing the model to its limits in incredible ways.
Here are 10 impressive examples of what it can do.
Llama-3 8B with a context length of over 1M
Developed by Gradient and sponsored by compute from Crusoe Energy, this model, called Llama-3 8B Gradient Instruct 1048k, extends LLama-3 8B’s context length from 8k to over 1048K. This model shows that SOTA LLMs can efficiently manage long contexts with minimal training by appropriately adjusting RoPE theta.
The model was trained progressively on increasing context lengths, drawing on techniques like NTK-aware interpolation and Ring Attention for efficient scaling. This approach allowed for a massive increase in training speed, making the model both powerful and efficient in handling extensive data.
We've been in the kitchen cooking 🔥 Excited to release the first @AIatMeta LLama-3 8B with a context length of over 1M on @huggingface – coming off of the 160K context length model we released on Friday!
— Gradient (@Gradient_AI_) April 29, 2024
A huge thank you to @CrusoeEnergy for sponsoring the compute. Let us know… pic.twitter.com/iZ9zcKzOc6
RAG App with Llama-3 running locally
You can build a RAG app with Llama-3 running locally on your computer (it’s 100% free and doesn’t require an internet connection).
The instructions include simple steps like installing the necessary Python Libraries, setting up the Streamlit App, creating Ollama embeddings and a vector store using Chroma, and setting up the RAG chain among other things.
Build a RAG app with Llama-3 running locally on your computer
— Shubham Saboo (@Saboo_Shubham_) May 2, 2024
(100% free and without internet):
Agri Vertical Dhenu 1.0 model fine-tuned on Llama3-8B
KissanAI’s Agri Vertical Dhenu1.0 model has been fine tuned on Llama3 8B for 150K instructions. It is India-focused and available for anyone to download, tinker and provide feedback.
Tool Calling Champion
Llama-3 70b on GroqInc is a tool-calling champion. The 70b model passed the task when given a query, was very fast, and had the best pricing. It’s also performing great at benchmarks and tests.
I found a new tool calling champion
— virat (@virattt) May 1, 2024
Llama3 70b on @GroqInc
Challenge: given user query, extract financial quarters and years.
Example: "How did revenue change between Q4 2023 and year before that?"
The 70b model:
• passed the task
• was very fast
• had best pricing
I… pic.twitter.com/q1UERftTMj
Lightning-fast Copilot in VSCode
You can connect @GroqInc with VSCode, unlocking the full potential of Llama-3 as your Copilot.
Just create your account on the Groq console, head to the ‘API Key’ menu and generate yours, download the CodeGPT extension from the VSCode marketplace. After this, open CodeGPT and select Groq as the provider, click on ‘Edit Connection’, paste your Groq API Key, and then click ‘Connect’.
That’s how you can connect Groq to VSCode and access all the models offered by this service.
Your own Copilot in VSCode lightning-fast thanks to Groq and Llama3
— Daniel San (@dani_avila7) May 2, 2024
In this thread 🧵, I'll guide you through connecting @GroqInc with VSCode, unlocking the full potential of Llama 3 (@AIatMeta) as your Copilot👇 pic.twitter.com/vy8lvqWXxQ
Llama-3 Function Calling
Llama-3 function calling works pretty well. Nous Research announced Hermes 2 Pro, which comes with Function Calling and Structured Output capabilities. The Llama-3 version now uses dedicated tokens for tool call parsing tags to make streaming function calls easier.
The model surpasses Llama-3 8B Instruct on AGIEval, GPT4All Suite, TruthfulQA, and BigBench.
Mind blowing 🤯 function calling by the new `Hermes 2 on Llama-3` by @Teknium1 @intrstllrninja running on @ollama
— Ashpreet Bedi (@ashpreetbedi) May 2, 2024
The last question asks it to do 3 different function calls and write an article. Just check that quality.
code: https://t.co/QLscvoY45i pic.twitter.com/xYIiOeLZea
TherapistAI, powered by Llama3-70B
TherapistAI.com now runs on Llama3-70B, which, according to the benchmarks, is almost as good as GPT-4. The Llama3-70B model significantly enhanced the app’s conversational capabilities, enabling a back-and-forth, ping-pong style interaction. The responses have become concise, direct, and highly focused on problem-solving.
With Llama-3, Therapist AI now actively engages by asking questions, which helps it understand better and address specific user needs. It also exhibits an impressive memory, allowing it to maintain context over longer conversations, thereby enhancing its ability to deliver relevant and actionable answers.
You can also use Llama-3 to build such applications. It ensures great performance and is less expensive than using ChatGPT 4, which is around $20 per month.
🧠 https://t.co/HDQWMHoORv now runs on Llama3-70B
— @levelsio (@levelsio) April 23, 2024
It's WAY better than Mixtral which is what I was running it on before, and it's now almost as good as GPT4 according to the benchmarks, but in my experience it's even better for therapy
Now with Llama3-70B:
– conversation like… https://t.co/dpOJlTxLTI pic.twitter.com/UfgZvxPvhU
AI Coding assistant with Llama 3
It’s time to give your productivity a boost by building an AI Coding assistant with Llama3.
To develop an AI coding assistant using Llama3, start by downloading Llama3 via Ollama, and then integrate a system message to enable it as a Python coding assistant. Next, install the Continue VSCode extension, connect it with my-python-assistant, and activate the tab-autocomplete feature to enhance coding efficiency.
Let's build an AI Coding assistant with Llama3 ↓🧵🦙
— Pau Labarta Bajo (@paulabartabajo_) April 29, 2024
Superfast Research Assistant using Llama 3
You can build a research assistant powered by Llama-3 models running on Groq. You can then take any complex topic, search the web for information about it, package it up, and send it to Llama-3 running on Groq. It will send back a proper research report.
Superfast Research Assistant using Llama3 on @GroqInc and @tavilyai
— Ashpreet Bedi (@ashpreetbedi) April 22, 2024
Build an Assistant that:
🔎 researches a complex topic
✍️ writes a report at 800 tokens/sec
Try it yourself: https://t.co/mRfeAIS6r7 pic.twitter.com/N3i8nT3cIx
Building RAG Capabilities for Accessing Private Data
Subtl.ai is building in-house RAG capabilities for accessing private data. Founded with the goal of democratizing access to private data for specific professional needs, the platform significantly improves efficiency by offering a 5x faster access to information. It does all this while maintaining data security through an AI that securely processes and recalls your data, allowing AI-enhanced access as well as data protection.
The company will be releasing their AI bot built on Llama-3 soon.