Ai Dreams Forum
Artificial Intelligence => AI News => Topic started by: MikeB on July 25, 2023, 08:47:36 am
-
https://ai.meta.com/llama/ (https://ai.meta.com/llama/)
https://www.llama2.ai/ (https://www.llama2.ai/)
This release includes model weights and starting code for pretrained and fine-tuned Llama language models — ranging from 7B to 70B parameters.
Llama 2 outperforms other open source language models on many external benchmarks, including reasoning, coding, proficiency, and knowledge tests.
Scores 80.2 on the Winogrande benchmark. Benchmarks are listed in the first URL.
Very similar politician/spokes-person style to both ChatGPT and Google Bard.
-
Have you tried the local AI LLMs like:
https://github.com/qwopqwop200/GPTQ-for-LLaMa
or
FreedomGPT:
https://www.youtube.com/watch?v=4IsDaM-uBLA&t=16s
Trying to get a gpu from google colab is tricky at best. I did look into another cloud gpu service which was charging, for the new H100s, 1.89/hr. But, they wanted 3 years up front and it's a 24/7 charge, that's like $50,000! You can get one on eBay for $30K (https://www.ebay.com/itm/195972662019?_trkparms=amclksrc%3DITM%26aid%3D1110006%26algo%3DHOMESPLICE.SIM%26ao%3D1%26asc%3D20200818143230%26meid%3D707ebde02fbb406296e3948ce72007d1%26pid%3D101224%26rk%3D5%26rkt%3D5%26sd%3D155598093022%26itm%3D195972662019%26pmt%3D0%26noa%3D1%26pg%3D4429486%26algv%3DDefaultOrganicWebV9BertRefreshRanker%26brand%3DNVIDIA&_trksid=p4429486.c101224.m-1) But the A100s have come down in price to as low as $5000 (https://www.ebay.com/itm/166227342736?hash=item26b3ebf590:g:gV8AAOSwaq1lAn3C&amdata=enc%3AAQAIAAAA4HTycDriEFi2iZtj2DQpR6kd%2FZdpmliAkdg7PI%2FScgLuI52TyujPk2NnQsiBk%2Bm%2FNUSk3cCSzx7UUaOOXbdR5lsaKiRxQFg71y0cccCwFB3kyQwwpghZxA8AcUbEc3byEHghqo1kUX8hwuNdJQ4oZzql%2F2VR5YJtYj5nWnqcjx6y7Fb7S1Ewgl3Wa%2BQKKwKXF54FrS6xQA%2BQShc7%2BziwuiU%2BCCIc9%2FxsSkS1LYdljYYNdFfu1S8tfpMNnJJNrFkT5Wq0W%2BKBUJbmozBPaOgsLrRTf5jk9CR6ZIRBD%2FxBifub%7Ctkp%3ABk9SR_bKme_bYg).
I recently purchased a RTX 3080 ti for $560 which replaced my M40. I like google colab and other cloud solutions but it becomes a real problem to get the high end gpus. If you're just needing inferencing a local gpu gets you in a better position. If you want to train those LLMs then the cloud solution is a good choice as long as you don't forget to signoff when your done, store your model on a cloud drive. But that's what I'm looking at now when it comes to customizing very Large LMs.
The LLMs can be converted to 4bit inferences which reduces the memory footprint quite a bit, so gpus like RTX 4090s look more interesting. Which now brings to mind some really interesting tools to literally build a virtual person:
https://www.youtube.com/watch?v=FKoy7bncHLs
https://www.youtube.com/watch?v=xFk_WU32igA
Now integrate the new AR glasses like XReal (https://www.amazon.com/Glasses-Micro-OLED-Augmented-iOS-Consoles-Compatible/dp/B0BF5LKP5Q/ref=sr_1_1_sspa?crid=36K3O6CW4IPCL&keywords=AR+glasses&qid=1695958527&sprefix=ar+glasse%2Caps%2C146&sr=8-1-spons&sp_csd=d2lkZ2V0TmFtZT1zcF9hdGY&psc=1) and you're looking at a pretty cool humaniod robot that you can take with you anywhere! :D
-
I assume that the local version of FreedomGPT is limited in some way with just 5GB of data?
-
I assume that the local version of FreedomGPT is limited in some way with just 5GB of data?
Try it, just download the version that will work with your OS.
-
I am reluctant to try it as I don't know what is in it.
-
Scan it with anti-virus. I tried and it worked fine.
-
I wonder if LLaMA2 works with llama-cpp-php by Bohuslav Šimek, and/or with llama.cpp by Georgi Gerganov ?
Reference: https://github.com/ggerganov/llama.cpp
Image source: Bohuslav Šimek <bohuslav@simek.si> Accessed via https://packagist.org/packages/kambo/llama-cpp-php MIT License. Permitted to use, copy, modify, merge publish, distribute, sublicense and sell copies of the software..
Image Modification: Zoomed in for easier reading by older people for discussion purposes only: There is a scroll bar at the bottom of the image to read the entire image.
-
Yes!
https://github.com/ggerganov/llama.cpp (https://github.com/ggerganov/llama.cpp)
Supported platforms:
...
LLaMA 2