SkySyrup

joined 2 years ago
MODERATOR OF
 

let’s see how low effort posts can be rule

CAPTION: A comment count totalling 196 circled in red

[–] SkySyrup@sh.itjust.works 2 points 2 years ago

I presume you have experience with these traps?

[–] SkySyrup@sh.itjust.works 1 points 2 years ago

UwU

I neeeeeeed it

 

I hope this isn’t a repost

[–] SkySyrup@sh.itjust.works 8 points 2 years ago* (last edited 2 years ago)

This looks exactly like an image that would accompany an SCP article lol

 

shamelessly stolen from nixCraft on mastodon

[–] SkySyrup@sh.itjust.works 1 points 2 years ago

the comment on that one are fun :)

[–] SkySyrup@sh.itjust.works 1 points 2 years ago

tem!! hOi!

sorry

 

I (was :( ) wearing a cute dress

 

a person holding a cat with the caption: It’s dangerous to go alone, take this

 

Content: creepy mark zuckerberg staring at camera with caption: This person tried to unlock your phone

[–] SkySyrup@sh.itjust.works 1 points 2 years ago

wow, the ceo of a ai company claims it can replace everyone!! It’s almost like it’s in his best interests to promise the moon to the shareholders to get as much money as possible!

[–] SkySyrup@sh.itjust.works 0 points 2 years ago* (last edited 2 years ago) (4 children)

You are correct. Hollywood will simply change up a couple things and then use the assets.

However, I‘m still undecided about how I think about whether generating AI art should count as Human-generated or not. On one hand, people can spend hours if not days or week perfecting a prompt with different tools like ControlNet, different promptstyles and etc. On the other hand, somebody comes up to midjourney, asks for a picture of a dragon wearing a T-Shirt and immediately gets an image that looks pretty decent. It’s probably not exactly what they wanted, but close enough, right? AI gets you 90% there what you want, and the other 10% is the super-hard part that takes forever. Anyway, sorry for dumping my though process from this comment chain on here xD

 
[–] SkySyrup@sh.itjust.works 1 points 2 years ago (1 children)

I'm using a Prusa Mini, works perfectly, no tuning, but I want to get an 0.6mm nozzle at some point.

[–] SkySyrup@sh.itjust.works 3 points 2 years ago (2 children)

I've seen the idea of a "reddit gold" being thrown around - eg you pay to get a post or comment a different color or effect and the money goes to the instance admins

[–] SkySyrup@sh.itjust.works 3 points 2 years ago

That's amazing! You should've scaled it to 110% and then shown it to your family xD

[–] SkySyrup@sh.itjust.works 2 points 2 years ago

I'm happy about the context improvements; I think it really helps!

 

The models after pruning can be used as is. Other methods require computationally expensive retraining or a weight update process.

Paper: https://arxiv.org/abs/2306.11695

Code: https://github.com/locuslab/wanda

Excerpts: The argument concerning the need for retraining and weight update does not fully capture the challenges of pruning LLMs. In this work, we address this challenge by introducing a straightforward and effective approach, termed Wanda (Pruning by Weights and activations). This technique successfully prunes LLMs to high degrees of sparsity without any need for modifying the remaining weights. Given a pretrained LLM, we compute our pruning metric from the initial to the final layers of the network. After pruning a preceding layer, the subsequent layer receives updated input activations, based on which its pruning metric will be computed. The sparse LLM after pruning is ready to use without further training or weight adjustment. We evaluate Wanda on the LLaMA model family, a series of Transformer language models at various parameter levels, often referred to as LLaMA-7B/13B/30B/65B. Without any weight update, Wanda outperforms the established pruning approach of magnitude pruning by a large margin. Our method also performs on par with or in most cases better than the prior reconstruction-based method SparseGPT. Note that as the model gets larger in size, the accuracy drop compared to the original dense model keeps getting smaller. For task-wise performance, we observe that there are certain tasks where our approach Wanda gives consistently better results across all LLaMA models, i.e. HellaSwag, ARC-c and OpenbookQA. We explore using parameter efficient fine-tuning (PEFT) techniques to recover performance of pruned LLM models. We use a popular PEFT method LoRA, which has been widely adopted for task specific fine-tuning of LLMs. However, here we are interested in recovering the performance loss of LLMs during pruning, thus we perform a more general “fine-tuning” where the pruned networks are trained with an autoregressive objective on C4 dataset. We enforce a limited computational budget (1 GPU and 5 hours). We find that we are able to restore performance of pruned LLaMA-7B (unstructured 50% sparsity) with a non-trivial amount, reducing zero-shot WikiText perplexity from 7.26 to 6.87. The additional parameters introduced by LoRA is only 0.06%, leaving the total sparsity level still at around 50% level. ​

NOTE: This text was largely copied from u/llamaShill

 

He's 15 years old now, and his ears really bother him, but he still brutally murders birds in our garden.

the fur on the sofa is from the other cats lol

 

This Community is new, but I plan to expand it and partially mirror posts from r/LocalLLaMA on Reddit.

 

Hi, you've found this ~~subreddit~~ Community, welcome!

This Community is intended to be a replacement for r/LocalLLaMA, because I think that we need to move beyond centralized Reddit in general (although obviously also the API thing).

I will moderate this Community for now, but if you want to help, you are very welcome, just contact me!

I will mirror or rewrite posts from r/LocalLLama for this Community for now, but maybe we could eventually all move to this Community (or any Community on Lemmy, seriously, I don't care about being mod or "owning" it).

view more: next ›