This you find around the ๐ŸŒ Internet
๐Ÿ” Web search with Google โ€ข Bing โ€ข DuckDuckGo โ€ข Marginalia โ€ข Reddit โ€ข Spotify โ€ข TikTok โ€ข Tootfinder โ€ข X โ€ข Yandex โ€ข Youtube โ€ข YTM
This you find in the ๐Ÿ›๏ธ Agora
๐Ÿค– AI assistant on 'candidate_sampling'
Generative AI services provided by Mistral AI. To save a generation into the Agora, for now please copy/paste into the document Stoa above.

Generating text...


๐Ÿ—ฃ๏ธ Stoas for [[candidate_sampling]]
A Stoa is a public space where people can meet and collaborate.
๐Ÿ“– Document at https://doc.anagora.org/candidate_sampling
๐Ÿ“น Meeting at https://framatalk.org/candidate_sampling
๐Ÿ“š Node [[candidate_sampling]]
๐Ÿ““ garden/KGBicheno/Artificial Intelligence/Introduction to AI/Week 3 - Introduction/Definitions/Candidate_Sampling.md by @KGBicheno

candidate sampling

Go back to the [[AI Glossary]]

A training-time optimization in which a probability is calculated for all the positive labels, using, for example, softmax, but only for a random sample of negative labels. For example, if we have an example labeled beagle and dog candidate sampling computes the predicted probabilities and corresponding loss terms for the beagle and dog class outputs in addition to a random subset of the remaining classes (cat, lollipop, fence). The idea is that the negative classes can learn from less frequent negative reinforcement as long as positive classes always get proper positive reinforcement, and this is indeed observed empirically. The motivation for candidate sampling is a computational efficiency win from not computing predictions for all negatives.

Loading pushes...

Rendering context...