
Overview

Coursera Plus Annual Sale:
All Certificates & Courses 50% Off!
Grab it
Learn how to deploy Llama 2 or Code Llama for enterprise use-cases in this 25-minute installation guide video from Trelis Research. Discover how to run Llama on a server you control and customize the user interface branding. Explore critical tools like text-generation-inference and chat-ui. Gain insights on selecting the best GPU, setting up an AWS server, installing Nvidia drivers, choosing the right Llama model, and configuring a Llama chatbot. The video covers topics such as setting up an API, adjusting temperature and length settings, and adding password protection. Access a private Github repo for detailed AWS and RunPod deployment instructions, support, and function calling inference scripts.
Syllabus
Deploy an Enterprise Language Model on AWS or on Google, Azure or Oracle.
How do I run Llama 2 on a server
What is the best GPU for running Llama 2
How to set up an AWS server for Llama 2
How to set up an api for llama with text-generation-inference
How to set up Nvidia drivers on an AWS server
What Llama model to choose?
Installing an app to run a Llama chatbot with chat-ui
How to select temperature and length for a language model
Adding password protection to your chat model and more
Taught by
Trelis Research