Welcome home, fellow Gator.

The Gator Nation's oldest and most active insider community
Join today!

Chinese Technology Continues To Improve And American Tech Should Be Concerned

Discussion in 'Too Hot for Swamp Gas' started by G8tas, Jan 27, 2025 at 9:10 AM.

  1. vaxcardinal

    vaxcardinal GC Hall of Fame

    8,183
    1,199
    2,543
    Apr 8, 2007
    Then probably just made up
     
  2. G8trGr8t

    G8trGr8t Premium Member

    33,407
    12,282
    3,693
    Aug 26, 2008
    Google Quantum AI team founder Hartmut Neven argued that the chip’s success supports the idea of quantum computation occurring in many parallel universes, aligning with interpretations of quantum mechanics that are based on a multiverse.

    Neven’s comments echo the theories of British physicist David Deutsch, who was among the first to suggest that quantum computation might involve parallel universes. Deutsch’s multiverse interpretation of quantum mechanics proposes that particles exist in multiple states simultaneously, a phenomenon that quantum computers leverage for their computational power.



    all I got out of it is that after going down several worm holes is there is no way that the computer they had set up could solve that math problem that fast unless it was working in conjunction with x number of other computers on different planes to help solve the problem..i think that was what it said, still not sure..

    quadrillion times the age of the universe..sounds like a long time...down to five mintues..not sure % efficiency increase but seems significant

    I want to know what the solution to the problem helps with..must be a BFD right?? or meaningless..either way...wow

    Google's 'Willow' quantum chip has solved a problem that would have taken the best supercomputer a quadrillion times the age of the universe to crack | Live Science


    Google scientists have created a new quantum processor that, in five minutes, cracked a problem that would have taken the world's best supercomputer 10 septillion years to solve. The breakthrough will allow quantum computers to become less error-prone the bigger they get, achieving a milestone that overcomes a decades-long obstacle.
     
  3. dingyibvs

    dingyibvs Premium Member

    2,079
    159
    293
    Apr 8, 2007
    My understanding is that the sanctions have only raised the cost a bit for China, they haven't really limited access. They have more than enough GPUs, and with Deepseek's breakthrough GPU demand will probably shift from the training side to the inference side, for which they don't really need Nvidia GPUs to do.

    As for Deepseek training using OpenAI's data, that is so much LOL, has to be irony at its richest. How dare you steal the data I just stole! :emoji_smile: It sounds like from the reports that Deepseek just used OpenAI's API, so probably no sensitive information is taken.
     
    • Fistbump/Thanks! Fistbump/Thanks! x 1
  4. G8trGr8t

    G8trGr8t Premium Member

    33,407
    12,282
    3,693
    Aug 26, 2008
    did you see the cerebrus chip performance running deepseek? it dusts the nvidia chip by a factor of 15x +/-. Cerebrus is aname to watch. worlds biggest chip, totally private company...something blackbox feel about it.

    Cerebras Launches World's Fastest DeepSeek R1 Distill Llama 70B Inference | Business Wire

    Cerebras Systems, the pioneer in accelerating generative AI, today announced record-breaking performance for DeepSeek-R1-Distill-Llama-70B inference, achieving more than 1,500 tokens per second – 57 times faster than GPU-based solutions. This unprecedented speed enables instant reasoning capabilities for one of the industry's most sophisticated open-weight models, running entirely on U.S.-based AI infrastructure with zero data retention.

    Powered by the Cerebras Wafer Scale Engine, the platform demonstrates dramatic real-world performance improvements. A standard coding prompt that takes 22 seconds on competitive platforms completes in just 1.5 seconds on Cerebras – a 15x improvement in time to result. This breakthrough enables practical deployment of sophisticated reasoning models that traditionally require extensive computation time.

    DeepSeek-R1-Distill-Llama-70B combines the advanced reasoning capabilities of DeepSeek's 671B parameter Mixture of Experts (MoE) model with Meta's widely-supported Llama architecture. Despite its efficient 70B parameter size, the model demonstrates superior performance on complex mathematics and coding tasks compared to larger models.

    "Security and privacy are paramount for enterprise AI deployment," continued Lupesko. "By processing all inference requests in U.S.-based data centers with zero data retention, we're ensuring that organizations can leverage cutting-edge AI capabilities while maintaining strict data governance standards. Data stays in the U.S. 100% of the time and belongs solely to the customer."