Code 0n Github: https://github.com/FlowiseAI/Flowise
The 6th incarnation of Elmer's blog
Code 0n Github: https://github.com/FlowiseAI/Flowise
Tutorial – train your own llama.cpp mini-ggml-model from scratch!
by u/Evening_Ad6637 in LocalLLaMA
Here I show how to train with llama.cpp your mini ggml model from scratch! these are currently very small models (20 mb when quantized) and I think this is more fore educational reasons (it helped me a lot to understand much more, when “create” an own model from.. nothing before. And it helps to understand the parameters and their effects much better)
Otherwise, these mini models could be good enough to be experts on very specific fields, like: only gives text in the style of someone. Like one model could speak like cartman from southpark, another could be a poem and you could implement these ‘person’ in your general chat or role play coversations as supporting roles or minor roles.. to make “group” chats, brainstormings, etc.
And: the discussions on github seems to be very promissing that we will soon be able to fine tune pre-trained big models like llama or vicuna and so on. espcially creating (q)lora adapters should be possible soon : )
this will be the next game changer i think (imagine your model could be finetuned in real time incrementally on top of its lora adapter and with your current conversation as the dataset – what awesome implications would this mean?)
EDIT:
You maybe need the training-script
— Tutorial – train your own llama.cpp mini-ggml-model from scratch!
Here’s 5 of the most valuable ways to convert unstructured text to structured data with natural language processing
Source: 5 Most Valuable Ways To Convert Unstructured Text To Structured Data | Width.ai
New large language models are publicly released almost every month. They are getting better and larger.
You may assume that these models can only be run on big clusters or in the cloud.
Fortunately, this is not the case. Recent versions of PyTorch propose several mechanisms that make the use of large language models relatively easy on a standard computer and without much engineering, thanks to the Hugging Face Accelerate package.
Source: Run Very Large Language Models on Your Computer | by Benjamin Marie | Towards AI
Unlock Your Second Brain with Streamlit and Hugging Face’s Free LLM Summarization: build a Python Webapp running on your PC.
Source: Mastering AI Summarization: Your Ultimate Productivity Hack
This uses a smaller language model tailored to text summarization. Maybe a good path for assessing student short answers and essays.
Developers can now fine-tune GPT-3 on their own data, creating a custom version tailored to their application. Customizing makes GPT-3 reliable for a wider variety of use cases and makes running the model cheaper and faster.
You can use an existing dataset of virtually any shape and size, or incrementally add data based on user feedback. With fine-tuning, one API customer was able to increase correct outputs from 83% to 95%. By adding new data from their product each week, another reduced error rates by 50%.
Now, a new study offers evidence supporting what […] has long observed: online social annotation helps students understand and construct knowledge around scholarly content, while at the same time building community.
https://www.insidehighered.com/news/2022/10/12/social-annotation-technology-helps-students-read-together
There is room for the use of social annotation tools in legal education especially as more teaching resources move online. Tools like Hypothesis would provide law students with ways to highlight and annotate online materials and share those as annotations with study groups, peers, and teachers.
Hypothesis is available in CALI Lawbooks, the online publishing platform that for CALI members
Canonical will provide free Ubuntu Pro on up to 5 machines for personal and small business uses. This provides additional security and support to extend the life of LTS releases. Details at https://9to5linux.com/canonical-launches-free-ubuntu-pro-subscriptions-for-everyone
Drupal 7 lives, but it lives only to die another day. So are you going to stay on Drupal 7? The answer, at some point, will need to be a resounding “no.”
Here’s a list of resources to get you started with DDEV and Drupal 9. As with setting up any new development (or production) environment there are a lot of moving parts and it take some time to get it all right. This list includes “HowTo” articles, tools, and documentation to get it all set up.
— After running ddev config and before running ddev start for the fist time use your favorite editor to edit .dev/config.yaml to the following:
name: d9-dev type: drupal9 docroot: web php_version: "8.1" webserver_type: apache-fpm router_http_port: "80" router_https_port: "443" xdebug_enabled: false additional_hostnames: [] additional_fqdns: [] mariadb_version: "" mysql_version: "8.0" nfs_mount_enabled: false mutagen_enabled: false use_dns_when_possible: true composer_version: "" web_environment: []
This will setup DDEV with MySQL 8, PHP 8.1, Drupal 9, and Apache. This matches the dev environment that CALI is using for D9. Check the DDEV docs for more possibilities.
— The DDEV install includes the latest phpmyadmin to help with mysql admin. It’s available in a local browser at <projectName>.ddev.site:8036. Use phpmyadmin to load a dump of the D9 dev database.
— Once WSL2 is setup, use Ubuntu 20.04 to host DDEV.
— DDEV includes git so that’s a good way to manage Drupal. In the CALI world use git to grab a copy of the current D9 code base.