Jump to content
Join the Unexplained Mysteries community today! It's free and setting up an account only takes a moment.
- Sign In or Create Account -

Large language models can be squeezed onto your phone — rather than needing 1000s of servers to run


Portre

Recommended Posts

A new algorithm, dubbed Calibration Aware Low precision Decomposition with Low Rank Adaptation (CALDERA), compresses the massive amounts of data needed to run a large language model (LLM) by trimming redundancies in the code and reducing the precision of its layers of information.

This leaner LLM performs with accuracy and nuance at slightly lower levels than the uncompressed version, scientists said in a study published May 24 to the preprint database arXiv, ahead of a presentation at the Conference on Neural Information Processing Systems (NeurIPS) in December.

Large language models can be squeezed onto your phone

  • Like 2
Link to comment
Share on other sites

 

Everyone (read Media, Education sector, Tech Corporations, Stock Market...) is saying AI is going to take over everything. 

Geeze, I hope that doesn't happen.

People are stupid enough already without having all their decisions made for them by their phone. 

Wonder when our equivalent of the Butlerian Jihad will happen.

Link to comment
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now
  • Recently Browsing   0 members

    • No registered users viewing this page.