llama.cpp on AWS EC2 under $2 #296
                  
                    
                      forgeda
                    
                  
                
                  started this conversation in
                Show and tell
              
            Replies: 0 comments
  
    Sign up for free
    to join this conversation on GitHub.
    Already have an account?
    Sign in to comment
  
        
    
Uh oh!
There was an error while loading. Please reload this page.
-
Prerequisites
Start a instance with t2.xlarge type hardware (4 core , 16GiB Memory), Ubuntu 22.04 AMI
SSH with PuTTY 0.76 above
Microsoft Remote Desktop Connection
1. Python Configuration
2. CMake Configuration
3. 7B Model Quantification and Inference with llama.cpp
4. Performance
5. Highlight
Beta Was this translation helpful? Give feedback.
All reactions