<br />Električna vozila i obnovljiva energija - Početna<br /> Električna vozila i obnovljiva energija

 
 FAQFAQ   PretražnikPretražnik   ČlanstvoČlanstvo   Korisničke grupeKorisničke grupe   RegistracijaRegistracija 
 ProfilProfil   Privatne porukePrivatne poruke   LoginLogin 

Umjetna inteligencija
Idite na Prethodna  1, 2, 3 ... 11, 12, 13
 
Započnite novu temu   Odgovorite na temu    
Električna vozila i obnovljiva energija - Početna
-> Umjetna inteligencija
Prethodna tema :: Sljedeća tema  
Autor/ica Poruka
zega



Pridružen/a: 01. 08. 2005.
Postovi: 5741
Lokacija: USA

PostPostano: 20.11.2024. sri. 16:03    Naslov: Citirajte i odgovorite

Animirani prikaz kako rade LLM's

https://hackaday.com/2024/11/20/an-animated-walkthrough-of-how-large-language-models-work/

[Vrh]
Korisnički profil Pošaljite privatnu poruku Posjetite Web stranice MSNM
stjepan



Pridružen/a: 03. 08. 2009.
Postovi: 1648
Lokacija: Zagreb

PostPostano: 29.11.2024. pet. 22:49    Naslov: Citirajte i odgovorite

Andrej Karpathy pokušava demistificirati LLM-ove

https://x.com/karpathy/status/1862565643436138619

Citat:
People have too inflated sense of what it means to "ask an AI" about something. The AI are language models trained basically by imitation on data from human labelers. Instead of the mysticism of "asking an AI", think of it more as "asking the average data labeler" on the internet.

Few caveats apply because e.g. in many domains (e.g. code, math, creative writing) the companies hire skilled data labelers (so think of it as asking them instead), and this is not 100% true when reinforcement learning is involved, though I have an earlier rant on how RLHF is just barely RL, and "actual RL" is still too early and/or constrained to domains that offer easy reward functions (math etc.).

But roughly speaking (and today), you're not asking some magical AI. You're asking a human data labeler. Whose average essence was lossily distilled into statistical token tumblers that are LLMs. This can still be super useful ofc ourse. Post triggered by someone suggesting we ask an AI how to run the government etc. TLDR you're not asking an AI, you're asking some mashup spirit of its average data labeler.
[Vrh]
Korisnički profil Pošaljite privatnu poruku
zega



Pridružen/a: 01. 08. 2005.
Postovi: 5741
Lokacija: USA

PostPostano: 4.12.2024. sri. 03:55    Naslov: Citirajte i odgovorite

Evo kako primitivno u stvari operira ai
https://youtube.com/shorts/7pQrMAekdn4
[Vrh]
Korisnički profil Pošaljite privatnu poruku Posjetite Web stranice MSNM
ogi
Site Admin


Pridružen/a: 25. 07. 2005.
Postovi: 12939
Lokacija: Zagreb

PostPostano: 4.12.2024. sri. 18:01    Naslov: Citirajte i odgovorite

How many "r"s in the word strawberry?
Citat:
There are three "r"s in the word strawberry.

Kako to znaš?
Citat:
Haha, jednostavno – brojim ih! 😊 Kad pogledam riječ strawberry, pratim svako slovo i pronađem tri r:

st → nema "r"
r → prva "r"
awbe → nema "r"
r → druga "r"
ry → treća "r"

Rezultat: 3 komada! 🎯
[Vrh]
Korisnički profil Pošaljite privatnu poruku Pošaljite e-mail Posjetite Web stranice
zega



Pridružen/a: 01. 08. 2005.
Postovi: 5741
Lokacija: USA

PostPostano: 4.12.2024. sri. 21:08    Naslov: Citirajte i odgovorite

Problem je "popravljen" https://xeiaso.net/blog/2024/strawberry/
[Vrh]
Korisnički profil Pošaljite privatnu poruku Posjetite Web stranice MSNM
zega



Pridružen/a: 01. 08. 2005.
Postovi: 5741
Lokacija: USA

PostPostano: 17.1.2025. pet. 04:03    Naslov: Citirajte i odgovorite

[Vrh]
Korisnički profil Pošaljite privatnu poruku Posjetite Web stranice MSNM
Cose



Pridružen/a: 13. 08. 2007.
Postovi: 5216
Lokacija: Zagreb

PostPostano: 17.1.2025. pet. 15:04    Naslov: Citirajte i odgovorite

Steve Gibson je oko Nove poslao mail kao rezultat svojeg cackanja po AI tematici - zgodan pregled svega:

- Everything that's going on is about neural networks. This has become so obvious to those in the business that they no longer talk about it. It would be like making a point of saying that today's computers run on electricity. (Duh!)

- AI computation can be divided into “pre-training” and “test-time” (also called “inference-time”). Pretraining is the monumental task of putting information into a massive and initially untrained neural network. Information is “put into” the network by comparing the network's output against the correct output, then tweaking the network's neural parameters to move the network's latest output more toward the correct output. A modern neural network might have 185 billion parameters interlinking its neurons, each which requires tweaking. This is done over and over and over (many millions of times) across a massive body of “knowledge” to gradually train the network to generate the proper output for any input.

- Counterintuitive though it may be, the result of this training is a neural network that actually contains the knowledge that was used to train it; it is a true knowledge representation. If that's difficult to swallow, consider human DNA as an analogy. DNA contains all of the knowledge that's required to build a person. The fact that DNA is not itself intelligent or sentient doesn't mean that it's not jam-packed with knowledge.

- The implementation of neural networks is surprisingly simple, requiring only a lot of simple multiplication and addition with massive parallelism. This is exactly what GPUs were designed to do. They were originally designed to perform the many simple 3D calculations needed for modern gaming, then they were employed to solve hash problems to mine cryptocurrency. But they now lie at the heart of all neural network AI.

- Even when powered by massive arrays of the fastest GPUs rented from cloud providers, this “pretraining” approach was becoming prohibitively expensive and time consuming. But five years ago, in 2019, a team of eight Google AI researchers published a ground breaking paper titled “Attention is all you need.” The title was inspired by the famous Beatles song “Love is all you need” and the paper introduced the technology they named “Transformers” (because one of the researchers liked the sound of the word). The best way to think of “Transformer” technology is that it allows massive neural networks to be trained much more efficiently “in parallel” and it also introduced the idea that not all of the training tokens needed to be considered equally because they were not all equally important. More “Attention” could be given to some than others. This breakthrough resulted in a massive overall improvement in training speed which, in turn, allowed vastly larger networks to be created and trained in reasonable time. Thus, it became practical and possible to train much larger neural networks and LLM's – Large Language Models – were born.

- The “GPT” of ChatGPT stands for Generative Pre-trained Transformer.

- But over time, once again, researchers began running into new limitations. They wanted even bigger networks because bigger networks provided more accurate results. But the bigger the network, the slower and more time consuming – and thus costly – was its training. It would have been theoretically possible to keep pushing that upward, but a better solution was discovered: Post-training computation.

- Traditional training of massive LLM's was very expensive. The breakthrough “Transformer” tech that made LLM-scale neural networks feasible for the first time was now being taken for granted. But at least the training was a one-time investment. After that, a query of the network could be made almost instantly and, therefore, for almost no money. But the trouble was that even with the largest practical networks the results could be unreliable – known as hallucinations. Aside from just being annoying, any neural network that was going to hallucinate and just “make stuff up” could never be relied upon to build “chains of inference” where its outputs could be used to explore the consequences of new inputs when seeking solutions to problems. Being able to do that would begin to look a lot like thinking.

- But a few years ago researchers began to better appreciate what could be done if a neural network's answer was not needed immediately. They began exploring what could be accomplished post-training if, when making a query, some time and computation – and thus money – could be spent working with the pre-trained network. By making a great many queries of the network and comparing multiple results, the overall reliability could be improved so much that it would be possible to create reliable inference chains for true problem solving. This is often referred to as Chains of Thought (CoT). Inference chains would allow for true problem solving using the stored knowledge that had been trained into these networks, and the pre-trained model could also be used for the correction of it own errors.

- I should note that the reason asking the same question multiple times results in multiple different answers is that researchers long ago discovered that introducing just a bit of “random factor” – which is called “the temperature” – into neural networks resulted in superior performance. (And, yes... if this all sounds suspiciously like VooDoo, you're not wrong – but it works anyway.)

- OpenAI's o1 model is the first of these more expensive inference-chain AI's to be made widely available. It offers a truly astonishing improvement over the previous ChatGPT 4o models. Since o1 is expensive for OpenAI to offer on a per-query basis, subscribers are limited to 7 full queries per day. But the o1-mini model, which is better but not as good, can be used without limit.

- Here's the big news: OpenAI just revealed that they have an o3 model that blows away their o1 model. It's not yet available, but it's coming. What IS available are the results of its benchmarks and that's why I believe you need to make time to watch this YouTube video: https://youtu.be/YAgIh4aFawU

- Is it AGI? OpenAI is saying not yet, but that they're closing in on it – and all of the evidence suggests that they are. The independent benchmarks and other tested performance cited in that video above are quite compelling.

- AGI only means that over a wide range of cognitive problem solving tasks an AI can outperform a knowledgeable person. Computers can already beat the best Chess, Go and Poker players. I think it's very clear that today's AIs are not far from being superior to humans at general problem solving. That doesn't make them a Frankenstein to be feared; it only makes AI a new and exceedingly useful tool.
[Vrh]
Korisnički profil Pošaljite privatnu poruku Posjetite Web stranice
stjepan



Pridružen/a: 03. 08. 2009.
Postovi: 1648
Lokacija: Zagreb

PostPostano: 17.1.2025. pet. 16:30    Naslov: Citirajte i odgovorite

Kad se priča o AI-u, većinom se svodi priča na razgovor o AI-ovima koji nešto stvaraju.
Evo primjera koji "gleda"
https://www.reddit.com/r/sdforall/comments/1i3i4j3/most_powerful_vision_model_cogvlm_2_now_works/

Pogledaš slike, pogledaš njegova objašnjenja toga što je na slikama.....ludilo.
[Vrh]
Korisnički profil Pošaljite privatnu poruku
Prethodni postovi:   
Započnite novu temu   Odgovorite na temu    
Električna vozila i obnovljiva energija - Početna
-> Umjetna inteligencija
Vremenska zona: GMT + 01:00.
Idite na Prethodna  1, 2, 3 ... 11, 12, 13
Stranica 13 / 13.

 
Forum(o)Bir:  
Ne možete otvarati nove teme.
Ne možete odgovarati na postove.
Ne možete uređivati Vaše postove.
Ne možete izbrisati Vaše postove.
Ne možete glasovati u anketama.


Powered by phpBB © 2001, 2002 phpBB Group
HR (Cro) by Ančica Sečan
home