• Wait till they figure out that China has also AI

    From Mild Shock@janburse@fastmail.fm to comp.lang.prolog on Thu Jan 2 20:03:01 2025
    From Newsgroup: comp.lang.prolog

    Hi,

    How it started:
    https://www.instagram.com/p/Cump3losObg

    How its going:
    https://9gag.com/gag/azx28eK

    Bye
    --- Synchronet 3.20a-Linux NewsLink 1.114
  • From Mild Shock@janburse@fastmail.fm to comp.lang.prolog on Tue Jan 28 01:34:32 2025
    From Newsgroup: comp.lang.prolog

    Hi,

    Wait till USA figures out there is a second
    competitor besides DeepSeek, its called Yi-Lightning:

    Yi-Lightning Technical Report
    https://arxiv.org/abs/2412.01253

    It was already discussed 2 months ago:

    Eric Schmidt DROPS BOMBSHELL: China DOMINATES AI! https://www.youtube.com/watch?v=ddWuEUjo4u4

    Bye

    Mild Shock schrieb:
    Hi,

    How it started:
    https://www.instagram.com/p/Cump3losObg

    How its going:
    https://9gag.com/gag/azx28eK

    Bye

    --- Synchronet 3.20c-Linux NewsLink 1.2
  • From Mild Shock@janburse@fastmail.fm to comp.lang.prolog on Fri Jan 31 16:23:53 2025
    From Newsgroup: comp.lang.prolog

    Hi,

    So how its going? DeepSeek embraced by many cloud
    providers, even by NVIDIA NIM itself.

    DeepSeek-R1 Now Live With NVIDIA NIM https://blogs.nvidia.com/blog/deepseek-r1-nim-microservice/

    What what are these models doing and how are they
    trained. Is Geoffrey Hinton our only AI God? There
    seems to be another slightly disputed AI God,

    S. Hochreiter, J. Schmidhuber. Long Short-Term Memory. Neural
    Computation, 9(8):1735-1780, 1997. https://people.idsia.ch/~juergen/deep-learning-history.html

    Bye

    P.S.: It allows a mechanistic view on our linguistic
    brain if the latent space is some semantic vectors?
    So that learning is a kind of control mechanism:

    Machine Learning Approach to Model Order Reduction
    of Nonlinear Systems via Autoencoder and LSTM Networks
    Thomas Simpson - 23 Sep 2021
    https://arxiv.org/abs/2109.11213

    Mild Shock schrieb:
    Hi,

    Wait till USA figures out there is a second
    competitor besides DeepSeek, its called Yi-Lightning:

    Yi-Lightning Technical Report
    https://arxiv.org/abs/2412.01253

    It was already discussed 2 months ago:

    Eric Schmidt DROPS BOMBSHELL: China DOMINATES AI! https://www.youtube.com/watch?v=ddWuEUjo4u4

    Bye

    Mild Shock schrieb:
    Hi,

    How it started:
    https://www.instagram.com/p/Cump3losObg

    How its going:
    https://9gag.com/gag/azx28eK

    Bye


    --- Synchronet 3.20c-Linux NewsLink 1.2
  • From Mild Shock@janburse@fastmail.fm to comp.lang.prolog on Fri Jan 31 23:49:08 2025
    From Newsgroup: comp.lang.prolog

    Hi,

    Please meet Luo Fuli:

    The 29-Year-Old Genius Behind DeepSeek’s AI Revolution https://www.youtube.com/watch?v=B2fxh4aoQ8Q

    I find this paper intersting, finally
    some say about fine tuning during pretraing:

    Raise a Child in Large Language Model
    13 Sep 2021 - Fuli Luo et al.
    https://arxiv.org/pdf/2109.05687

    Bye

    Mild Shock schrieb:
    Hi,

    So how its going? DeepSeek embraced by many cloud
    providers, even by NVIDIA NIM itself.

    DeepSeek-R1 Now Live With NVIDIA NIM https://blogs.nvidia.com/blog/deepseek-r1-nim-microservice/

    What what are these models doing and how are they
    trained. Is Geoffrey Hinton our only AI God? There
    seems to be another slightly disputed AI God,

    S. Hochreiter, J. Schmidhuber. Long Short-Term Memory. Neural
    Computation, 9(8):1735-1780, 1997. https://people.idsia.ch/~juergen/deep-learning-history.html

    Bye

    P.S.: It allows a mechanistic view on our linguistic
    brain if the latent space is some semantic vectors?
    So that learning is a kind of control mechanism:

    Machine Learning Approach to Model Order Reduction
    of Nonlinear Systems via Autoencoder and LSTM Networks
    Thomas Simpson - 23 Sep 2021
    https://arxiv.org/abs/2109.11213

    Mild Shock schrieb:
    Hi,

    Wait till USA figures out there is a second
    competitor besides DeepSeek, its called Yi-Lightning:

    Yi-Lightning Technical Report
    https://arxiv.org/abs/2412.01253

    It was already discussed 2 months ago:

    Eric Schmidt DROPS BOMBSHELL: China DOMINATES AI!
    https://www.youtube.com/watch?v=ddWuEUjo4u4

    Bye

    Mild Shock schrieb:
    Hi,

    How it started:
    https://www.instagram.com/p/Cump3losObg

    How its going:
    https://9gag.com/gag/azx28eK

    Bye



    --- Synchronet 3.20c-Linux NewsLink 1.2
  • From Mild Shock@janburse@fastmail.fm to comp.lang.prolog on Sun Feb 2 16:35:13 2025
    From Newsgroup: comp.lang.prolog

    Hi,

    Remember Mira Murati? I do not either. Now
    China has this secret weapon Mata Hari:

    DeepSeek Principal Researcher: Luo Fuli. https://x.com/jenzhuscott/status/1885542484543234150

    If you’re wondering how China keeps stealing all
    our technology secrets so easily, it’s this. They
    send over girls that look like this.

    White nerds are completely defenseless against
    this. This is a marketing bomb if they use it.
    They already have hearts and minds, they can

    carpet bomb the little remaining goodwill
    towards the CS models.

    Bye

    P.S.: But Jürgen Schmidhuber still avoids
    the word bilinear? Why? Can one not come to
    the conclusion that we don't need some signmoids

    and can directly quantizise a model during learning?

    DeepSeek uses a distilled chain of thought system? https://x.com/SchmidhuberAI/status/1885357355938046382

    Mild Shock schrieb:
    Hi,

    How it started:
    https://www.instagram.com/p/Cump3losObg

    How its going:
    https://9gag.com/gag/azx28eK

    Bye

    --- Synchronet 3.20c-Linux NewsLink 1.2
  • From Mild Shock@janburse@fastmail.fm to comp.lang.prolog on Tue Feb 4 10:03:17 2025
    From Newsgroup: comp.lang.prolog

    Hi,

    Because of the wide availability of Machine Learning
    via Python libraries , the whole world (at least China)
    has become a big Petri Dish that is experimenting with

    new strategies to evolve brains on the computer.
    Recent discovery seems to be Group Preference Optimization.
    This is when you make the chat bot, detect and react

    differently to different groups of people. It seems to
    work on the "policy level". I don't understand it yet
    completely. But chat bots can then evolve and use

    multiple policies automatically:

    Group Preference Optimization
    https://arxiv.org/abs/2310.11523

    DeepSeekMath: Pushing the Limits
    https://arxiv.org/abs/2402.03300

    Now it seems that it is also at the core of DeepSeekMath,
    what is possibly detected is not group of people, but
    mathematical topics, so that in the end it excells.

    When unsupervised learning is used groups or math
    topics might be found from data, through a form of
    abduction.

    Bye

    Mild Shock schrieb:
    Hi,

    Wait till USA figures out there is a second
    competitor besides DeepSeek, its called Yi-Lightning:

    Yi-Lightning Technical Report
    https://arxiv.org/abs/2412.01253

    It was already discussed 2 months ago:

    Eric Schmidt DROPS BOMBSHELL: China DOMINATES AI! https://www.youtube.com/watch?v=ddWuEUjo4u4

    Bye

    Mild Shock schrieb:
    Hi,

    How it started:
    https://www.instagram.com/p/Cump3losObg

    How its going:
    https://9gag.com/gag/azx28eK

    Bye


    --- Synchronet 3.20c-Linux NewsLink 1.2
  • From Mild Shock@janburse@fastmail.fm to comp.lang.prolog on Sat Feb 8 12:57:10 2025
    From Newsgroup: comp.lang.prolog

    Hi,

    I try to motivate a Biology Teacher already for a while to
    replicate the below grokking experiment. But I have my
    own worries, why bother with the blackbox of what a

    machine learning method has learnt?

    Simple PyTorch Implementation of "Grokking" https://github.com/teddykoker/grokking

    Well its not correct to say that the learnt model is a black box.
    The training data was somehow a black box, but the resulting
    model is a white box, you can inspect it.

    This gives rise to a totally new scientific profession of
    full time artificial intelligence model gazers. And it is
    aprils fools day all year long:

    Language Models Use Trigonometry to Do Addition https://arxiv.org/abs/2502.00873

    Have Fun!

    Bye

    Mild Shock schrieb:
    Hi,

    Because of the wide availability of Machine Learning
    via Python libraries , the whole world (at least China)
    has become a big Petri Dish that is experimenting with

    new strategies to evolve brains on the computer.
    Recent discovery seems to be Group Preference Optimization.
    This is when you make the chat bot, detect and react

    differently to different groups of people. It seems to
    work on the "policy level". I don't understand it yet
    completely. But chat bots can then evolve and use

    multiple policies automatically:

    Group Preference Optimization
    https://arxiv.org/abs/2310.11523

    DeepSeekMath: Pushing the Limits
    https://arxiv.org/abs/2402.03300

    Now it seems that it is also at the core of DeepSeekMath,
    what is possibly detected is not group of people, but
    mathematical topics, so that in the end it excells.

    When unsupervised learning is used groups or math
    topics might be found from data, through a form of
    abduction.

    Bye

    Mild Shock schrieb:
    Hi,

    Wait till USA figures out there is a second
    competitor besides DeepSeek, its called Yi-Lightning:

    Yi-Lightning Technical Report
    https://arxiv.org/abs/2412.01253

    It was already discussed 2 months ago:

    Eric Schmidt DROPS BOMBSHELL: China DOMINATES AI!
    https://www.youtube.com/watch?v=ddWuEUjo4u4

    Bye

    Mild Shock schrieb:
    Hi,

    How it started:
    https://www.instagram.com/p/Cump3losObg

    How its going:
    https://9gag.com/gag/azx28eK

    Bye



    --- Synchronet 3.20c-Linux NewsLink 1.2