• 4 Posts
  • 25 Comments
Joined 4 months ago
cake
Cake day: September 12th, 2025

help-circle




  • Hmm, if you’re asking about me specifically, the LLMs I have on my PC are small and vastly outclassed by models hosted online. I don’t have a specific use case for them other than personal amusement and familiarising myself with the technology, and I don’t gain much from using them either.

    As for how China specifically is developing this technology, the main positive aspect is that a majority of LLMs released by Chinese firms and research groups have the model weights open under free software licenses, so everyone can download and tweak them.

    Certainly, I do not think that Chinese tech firms have the people’s interests at heart any more than other companies, but given that a push for open source AI is explicitly part of the 14th 5 year plan, I think it’s pretty clear the CPC is aware of the exploitative potential of these technologies, and is actively working to minimise the risk.









  • Most AI platforms use massive models with trillions of parameters that activate all their computational power for every single query.

    The first part is probably right, frontier models are likely around a trillion parameters total, though we don’t know for sure, but I don’t think the second part is correct. It’s almost certain the big proprietary models, like all the recent large open models, use a Mixture-of-Experts system like Thuara, because it’s just cheaper to train and run.

    While traditional web research might take you 10 minutes of clicking through pages and consuming cookies from major search engines, Thaura uses a fraction of the energy and provides the same information.

    This part is pretty misleading. It is very unclear how much an LLM query compares to a search in terms of energy use, but even assuming they’re about equal (most estimates put LLM queries higher), the LLM also has to do their own web searches to find the information, if you’re using it for research purposes, so that point is fairly moot. Also the “consuming cookies” part isn’t really an energy problem, but a privacy problem, so I’m not sure why it’s used in this context.

    Thaura uses a “mixture of expert” models with 100 billion total parameters, but only activates 12 billion per query.

    Going to the actual website, it does credit the “GLM-4.5 Air architecture”, but the article doesn’t mention GLM, or the company behind it (Z.ai) at all. Given that this is likely a finetune of the GLM model that was freely released, it feels weird how the Thaura team seem reluctant to give credit to the Z.ai team.

    These companies are often controlled by US-based corporations whose political stance supports occupation, apartheid, and Western hegemony - not human rights or global justice.

    Reading below and also looking at their website, the hosting and inference is done by US firms (DigitalOcean, TogetherAI) in datacenters hosted in the EU. That’s not inherently bad from a privacy standpoint due to encryption, but it does feel disjointed that they are railing against US firms and western hegemony while simultaneously using their services for Thaura.

    While I don’t think the Thaura team had bad intentions in fine-tuning their model and building the service, I feel that this is a pretty misleading article that also doesn’t give any significant details on Thaura, like it’s performance. They also haven’t given back to the community by releasing their model weights, despite building on an open model themselves. Personally, I think it’s better to stick to Z.ai, Qwen, Deepseek, etc, who actually release their models to the community and pretrain their models themselves.







  • Apparently they’re currently spending 6.3% of GDP on defense this year. In comparison, the US spent 9.4% during the height of the Vietnam war, the CIA estimates the USSR spent around 10-15% through the late 70s and 80s, and the Nazis spent 25% in 1939 and 75% in 1944.

    I haven’t really done a rigorous check on these numbers, and all of them are from western sources so take them with as much salt as you’d like, but it’s pretty clear Russia is still far from a true war economy and have significant room to ramp up further if they decide they need to. Also, they have a large trade surplus with both China and India, so they have a consistent source of revenue.


  • Anecdotally I do feel like there is, to a degree, more of a pro-China undercurrent to mainstream-ish discourse than there was before. Outright celebration and respect of China and the accomplishments of their people is still rare, but I’m seeing a lot of “Well, at least they’re doing [insert X] better than us.” Even if they just have to add a “But authoritarian!” somewhere.

    I think there’s both a push and pull effect, Trump is an obvious push factor, but there is also a rising disillusionment of western governments’ ability to deliver a better future for their people. The pull factors meanwhile are China’s visible achievements in building a prosperous society, and the rise in Chinese cultural exports like video games, movies, and technology.

    The main challenge for us is to maintain this sentiment after the Democrats take over and the liberals all fall in line. I think that emphasising what China does well, not just what China does better than Trump, is a good start. We need to erode the omnipresent idea that only liberal electoral systems can produce good outcomes for their people, and China makes it pretty easy for us to point to as a counterexample.