You Can Now Run a GPT-3 Level AI Model On Your Laptop, Phone, and Raspberry Pi

On Friday, a software developer named Georgi Gerganov created a tool called “llama.cpp” that can run Meta’s new GPT-3-class AI large language model, LLaMA, locally on a Mac laptop. Soon thereafter, people worked out how to run LLaMA on Windows as well. Then someone showed it running on a Pixel 6 phone, and next came a Raspberry Pi (albeit running very slowly). If this keeps up, we may be looking at a pocket-sized ChatGPT competitor before we know it. […]

Typically, running GPT-3 requires several datacenter-class A100 GPUs (also, the weights for GPT-3 are not public), but LLaMA made waves because it could run on a single beefy consumer GPU. And now, with optimizations that reduce the model size using a technique called quantization, LLaMA can run on an M1 Mac or a lesser Nvidia consumer GPU. After obtaining the LLaMA weights ourselves, we followed [independent AI researcher Simon Willison’s] instructions and got the 7B parameter version running on an M1 Macbook Air, and it runs at a reasonable rate of speed. You call it as a script on the command line with a prompt, and LLaMA does its best to complete it in a reasonable way.

There’s still the question of how much the quantization affects the quality of the output. In our tests, LLaMA 7B trimmed down to 4-bit quantization was very impressive for running on a MacBook Air — but still not on par with what you might expect from ChatGPT. It’s entirely possible that better prompting techniques might generate better results. Also, optimizations and fine-tunings come quickly when everyone has their hands on the code and the weights — even though LLaMA is still saddled with some fairly restrictive terms of use. The release of Alpaca today by Stanford proves that fine tuning (additional training with a specific goal in mind) can improve performance, and it’s still early days after LLaMA’s release. A step-by-step instruction guide for running LLaMA on a Mac can be found here (Warning: it’s fairly technical).

Source: You Can Now Run a GPT-3 Level AI Model On Your Laptop, Phone, and Raspberry Pi – Slashdot

NL Museum Temporarily Swapped Girl with Pearl Earring With AI Art and Luddites didn’t like it – sounds like art pushing boundries

A Netherlands museum is facing criticism for selecting an AI-generated piece of art to temporarily take the place of the renowned Girl with a Pearl Earring painting. The artwork was created by Johannes Vermeer in 1665 and is usually located at the Mauritshuis Museum but is on loan at the Rijksmuseum in Amsterdam until June 4.

In the interim, the Mauritshuis Museum held a competition for local artists to submit their own versions of the Girl with a Pearl Earring painting and said it would select one of the submissions to take Vermeer’s place until the painting is returned. While the competition may have seemed like a straightforward and exciting process, when the museum selected an AI-generated piece of art showing the girl with more structured and sharp outlines and glowing earrings, the art community erupted with complaints.

Of the roughly 3,480 artworks submitted, Berlin-based artist, Julian van Dieken, was one of five winners selected, and whose so-called painting is receiving backlash from artists and lovers of the painting.

[…]

When asked for comment, the Mauritshuis Museum directed Gizmodo to a statement on its website which said they did not choose the winners by looking at what was the “most beautiful” or “best” submission. “For us, the starting point has always been that the maker has been inspired by Johannes Vermeer’s world-famous painting. And that can be in the most diverse ways in image or technique.”

Source: A Museum Temporarily Swapped a Masterpiece With AI Art and Chaos Ensued

Sounds like art doing what art should be doing – pushing culture and perceptions. Making people think. Just a shame the village idiot squad angle is pushed by Gizmodo. Well done the Mauritshuis!

Google adds ChatGPT AI rival PaLM to Docs, Gmail, API and more

Google has promised to offer API-level access to its large language model PaLM so that developers can build it into their apps and workflows, and thus make the ChatGPT-like text-emitting tech available to world-plus-dog.

The web giant is also threatening to bake the model’s content-generating capabilities into Google Docs, Gmail, and more.

[…]

On Tuesday, Google unveiled its PaLM API, opening up its text-generating large language model to developers looking to boost their applications with auto-generated machine-made writing and other stuff. It’s capable of summarizing and classifying text, acting as a support chat bot that interacts with folks on behalf of your organization, and other things, just like the other APIs out there from OpenAI, Cohere, and AI21 Labs.

[…]

PaLM API also comes with MakerSuite, a tool that allows developers to experiment with the model by trying different prompts to fine-tune the model’s output. These software services are available to a select few, however, for the moment: Google is gradually rolling them out.

The internet goliath promises that general users can look forward to eventually being able to automatically generate email drafts and replies, as well as summarize text. Images, audio, and video created using the AI engine will be available to add to Slides, whilst better autocomplete is coming to Sheets. New backgrounds and note-generating features are also coming to Meet.

[…]

Source: Google adds ChatGPT AI rival to Docs, Gmail, API and more • The Register

Anthropic launches Claude, a chatbot to rival OpenAI’s ChatGPT

Anthropic, a startup co-founded by ex-OpenAI employees, today launched something of a rival to the viral sensation ChatGPT.

Called Claude, Anthropic’s AI — a chatbot — can be instructed to perform a range of tasks, including searching across documents, summarizing, writing and coding, and answering questions about particular topics. In these ways, it’s similar to OpenAI’s ChatGPT. But Anthropic makes the case that Claude is “much less likely to produce harmful outputs,” “easier to converse with” and “more steerable.”

Organizations can request access. Pricing has yet to be detailed.

[…]

Following a closed beta late last year, Anthropic has been quietly testing Claude with launch partners, including Robin AI, AssemblyAI, Notion, Quora and DuckDuckGo. Two versions are available as of this morning via an API, Claude and a faster, less costly derivative called Claude Instant.

[…]

closed beta. Claude is reportedly worse at math and a poorer programmer than ChatGPT. And it hallucinates, inventing a name for a chemical that doesn’t exist, for example, and providing dubious instructions for producing weapons-grade uranium.

It’s also possible to get around Claude’s built-in safety features via clever prompting, as is the case with ChatGPT. One user in the beta was able to get Claude to describe how to make meth at home.

“The challenge is making models that both never hallucinate but are still useful — you can get into a tough situation where the model figures a good way to never lie is to never say anything at all, so there’s a tradeoff there that we’re working on,” the Anthropic spokesperson said. “We’ve also made progress on reducing hallucinations, but there is more to do.”

Anthropic’s other plans include letting developers customize Claude’s constitutional principles to their own needs. Customer acquisition is another focus, unsurprisingly — Anthropic sees its core users as “startups making bold technological bets” in addition to “larger, more established enterprises.”

[…]

The company has substantial outside backing, including a $580 million tranche from a group of investors including disgraced FTX founder Sam Bankman-Fried, Caroline Ellison, Jim McClave, Nishad Singh, Jaan Tallinn and the Center for Emerging Risk Research.

Most recently, Google pledged $300 million in Anthropic for a 10% stake in the startup. Under the terms of the deal, which was first reported by the Financial Times, Anthropic agreed to make Google Cloud its “preferred cloud provider” with the companies “co-develop[ing] AI computing systems.”

Source: Anthropic launches Claude, a chatbot to rival OpenAI’s ChatGPT | TechCrunch

OpenAI Levels Up, commercialises more With Newly Released GPT-4

[…]

On Tuesday, the company unveiled GPT-4, an update to its advanced AI system that’s meant to generate natural-sounding language in response to user input. The company claimed GPT-4 is more accurate and more capable of solving problems. It even inferred that ChatGPT performs better than most humans can on complicated tests. OpenAI said GPT-4 scores in the 90th percentile of the Uniform Bar Exam and the 99th percentile of the Biology Olympiad. GPT-3, the company’s previous version, scored 10th and 31st on those tests, respectively.

The new system is now capable of handling over 25,000 words of text, according to the company. GPT-3 was only capable of handling 2,048 linguistic tokens, or 1,500 words at a time. This should allow for “more long-from content creation.” That’s not to say some folks haven’t tried writing entire novels with earlier versions of the LLM, but this new version could allow text to remain much more cohesive.

Those who have been hanging on OpenAI’s every word have been long anticipating the release of GPT-4, the latest edition of the company’s large language model. OpenAI said it spent six months modifying its LLM to make it 82% less likely to respond to requests for “disallowed content” and 40% more likely to produce factual responses than previous versions. Of course, we don’t have access to OpenAI’s internal data that might show how often GPT-3 was liable to lie or showcase banned content. Few people outside OpenAI have been able to take the new system on a test run, so all these claims could very well just be mere puffery.

Folks looking to get access to GPT-4 either has to be one of the select few companies given early access, or join a waitlist for the GPT-4 API or be one of the lucky few selected ChatGPT Plus subscribers.

The new system also includes the ability to accept images as inputs, allowing the system to generate captions, or provide analyses of an image. The company used the example of an image with a few ingredients, and the system provided some examples for what food those ingredients could create. OpenAI CEO Sam Altman wrote on Twitter that the company was “previewing” its visual inputs but it will “need some time to mitigate the safety challenges.”

What else is GPT-4 good at?

In a Tuesday livestream, OpenAI showed off a few capabilities of GPT-4, though the company constantly had to remind folks to not explicitly trust everything the AI produces.

In the livestream, OpenAI President Greg Brockman showed how the system can complete relatively inane tasks, like summarizing an article in one sentence where every word starts with the same letter. He then showed how users can instill the system with new information for it to parse, adding parameters to make the AI more aware of its role.

The company co-founder said the system is relatively slow, especially when completing complex tasks, though it wouldn’t take more than a few minutes to finish up requests. In one instance, Brockman made the AI create code for an AI-based Discord bot. He constantly iterated on the requests, even inputting error messages into GPT-4 until it managed to craft what was asked. He also put in U.S. tax code to finalize some tax info for an imaginary couple.

All the while, Brockman kept reiterating that people should not “run untrusted code from humans or AI,” and that people shouldn’t implicitly trust the AI to do their taxes. Of course, that won’t stop people from doing exactly that, depending on how capable public models of this AI end up being. It relates to the very real risk of running these AI models in professional settings, even when there’s only a small chance of AI error.

“It’s not perfect, but neither are you,” Brockman said.

OpenAI is getting even more companies hooked on AI

OpenAI has apparently leveraged its recently-announced multi-billion dollar arrangement with Microsoft to train GPT-4 on Microsoft Azure supercomputers. Altman said this latest version of the company’s LLM is “more creative than previous models, it hallucinates significantly less, and it is less biased.” Still, he said the company was inviting more outside groups to evaluate GPT-4 and offer feedback.

Of course, that’s not to say the system isn’t already been put into use by several companies. Language learning app Duolingo announced Tuesday afternoon that it was implementing a “Duolingo Max” premium subscription tier. The app has new features powered by GPT-4 that lets AI offer “context-specific explanations” for why users made a mistake. It also lets users practice conversations with the AI chatbot, meaning that damn annoying owl can now react to your language flubs in real time.

Because that’s what this is really about, getting more companies to pay to access OpenAI’s APIs. Altman mentioned the new system will have even more customization of behavior, which will further allow developers to fine-tune AI for specific purposes. Other customers of GPT-4 include the likes of Morgan Stanley, Khan Academy, and the Icelandic government. The U.S. Chamber of Commerce recently said in 10 years, virtually every company and government entity will be up on this AI tech.

Though the company still said GPT-4 has “many known limitations” including social biases, hallucinations, and adversarial prompts. Even if the new system is better than before, there’s still plenty of room for the AI to be abused. Some ChatGPT users have flooded open submission sections for at least one popular fiction magazine. Now that GPT-4 can write even longer, It’s likely we’ll see even more long-form AI-generated content flooding the internet.

Source: OpenAI Levels Up With Newly Released GPT-4

OpenAI was supposed to be all about open source and stuff, but with this definitely being about increasing (paid) API access, it’s looking more and more like a massive money grab. Not really surprising but a real shame.