Tobberone

joined 1 year ago
[–] Tobberone@lemm.ee 1 points 1 month ago

There must be. Recall and info sec is mutually excluding by definition!

[–] Tobberone@lemm.ee 2 points 1 month ago

I'm just in the beginning, but my plan is to use it to evaluate policy docs. There is so much context to keep up with, so any way to load more context into the analysis will be helpful. Learning how to add excel information in the analysis will also be a big step forward.

I will have to check out Mistral:) So far Qwen2.5 14B has been the best at providing analysis of my test scenario. But i guess an even higher parameter model will have its advantages.

[–] Tobberone@lemm.ee 2 points 1 month ago (2 children)

Thank you! Very useful. I am, again, surprised how a better way of asking questions affects the answers almost as much as using a better model.

[–] Tobberone@lemm.ee 1 points 1 month ago (1 children)

I need to look into flash attention! And if i understand you correctly a larger model of llama3.1 would be better prepared to handle a larger context window than a smaller llama3.1 model?

[–] Tobberone@lemm.ee 1 points 1 month ago (1 children)

Thanks! I actually picked up the concept of context window, and from there how to create a modelfile, through one of the links provided earlier and it has made a huge difference. In your experience, would a small model like llama3.2 with a bigger context window be able to provide the same output as a big modem L, like qwen2.5:14b, with a more limited window? The bigger window obviously allow more data to be taken into account, but how does the model size compare?

[–] Tobberone@lemm.ee 2 points 1 month ago (1 children)

Thank you for your detailed answer:) it's 20 years and 2 kids since I last tried my hand at reading code, but I'm doing my best to catch up😊 Context window is a concept I picked up from your links which has provided me much help!

[–] Tobberone@lemm.ee 1 points 1 month ago (3 children)

The problem I keep running into with that approach is that only the last page is actually summarised and some of the texts are... Longer.

[–] Tobberone@lemm.ee 5 points 1 month ago (15 children)

Do you know of any nifty resources on how to create RAGs using ollama/webui? (Or even fine-tuning?). I've tried to set it up, but the documents provided doesn't seem to be analysed properly.

I'm trying to get the LLM into reading/summarising a certain type of (wordy) files, and it seems the query prompt is limited to about 6k characters.

[–] Tobberone@lemm.ee -1 points 1 month ago

Why ruin goat cheese and bacon like that?

[–] Tobberone@lemm.ee 5 points 2 months ago (1 children)

Had blocking news and access to information been in the cards, as you describe, there would be another discussion. This is not it. The closest this comes is to block a linkaggregator. One that has been deemed to violate the laws in its area of business and being reluctant to take steps to rectify the situation.

This being the supreme court doing it does bring up the question of democratic decision making, which famiously has been proven by other countries recently. Although they also gave their president the power to remove themselves from office, if I've understood that particular debacle correctly.

[–] Tobberone@lemm.ee 0 points 3 months ago (7 children)

That's not a straight line, although it is possible to follow without changing direction😊

[–] Tobberone@lemm.ee 0 points 4 months ago

Well... Its built on statistics and statistical inference will return to the mean eventually. If all it ever gets to train on is closer and closer to the mean, there will be nothing left to work with. It will all be the average...

view more: next ›