Unexpected poetry. Is there a reason why crayfish would be served in this context?
https://mychinesehomekitchen.com/2022/06/24/chinese-style-sp...
So yes, I read it the same way you do: “They made us work weekends, but at least they’d order us in some pizzas.”
(…and if you’re in the US, you can have them air-freighted live to you, and a crawfish boil is an easy and darn festive thing to do in the summer. If you’re put off by the crustacean staring back at you, and you have access to a kitchen that operates in a Louisianan style, you might be able to find a “Cajun Popcorn” of the tails seasoned, battered, and fried. Or maybe one of the enormous number of “seafood boil” restaurants that have opened in the US in recent years.)
(I feel like those establishments came on quickly, that I notice them mainly in spaces formerly occupied by American-Chinese restaurants, and that it’s felt like a nationwide phenomenon… I suspect there’s a story there for an enterprising young investigative nonfiction writer sort.)
> Welcome to OML 1.0: Fingerprinting. This repository houses the tooling for generating and embedding secret fingerprints into LLMs through fine-tuning to enable identification of LLM ownership and protection against unauthorized use.
I started doing that once LLM provided me with a solution to a problem that was quite elegant, but was not implemented in the particular project. Turns out it learned it from GitHub issues post that described how particular problem could be tackled, but PR never actually got in.
实际上,对于后续训了很久很久的这个模型,Honestagi能够分析出这个量级的相似性我已经很诧异了,因为这个模型为了续训洗参数,所付出的算力甚至早就足够从头训一个同档位的模型了。听同事说他们为了洗掉千问的水印,采取了不少办法,甚至包括故意训了脏数据。这也为学术界研究模型血缘提供了一个前所未有的特殊模范吧。以后新的血缘方法提出可以拿出来溜溜。
In fact, I'm surprised that HonestAGI's analysis could show this level of similarity for this model that had been post-trained for a long time, because the computing power used to train-wash the parameters of this model was enough to train a model of the same size from scratch. I heard from my colleagues that they took many measures to wash off Qwen's watermark, even deliberately training on dirty data. This also provides an unprecedented case study for the academic community studying model lineage. If a new lineage method is put forward in the future, you can take it for a spin.
Very funny that the chinese even do this to each other; equal-opportunity cheats.
This article on the topic is a good explainer, https://aeon.co/essays/why-in-china-and-japan-a-copy-is-just... , but it's a thoroughly studied phenomenon.
To get real proof we created a new program that only existed on a single machine, and then added a signature for that application. This way there could be no claim that they independently added something to their database, as the program was not malware and literally impossible to actually find in the wild. Once they added it to their database we made a blog post and the issue got a lot of attention.
https://forums.malwarebytes.com/topic/29681-iobit-steals-mal...
I hope you were not the one that decided to uninstall the product, you need to download a support utility... :-)
The issues in this report are really about: 1. Lies about Huawei's capabilities to the country (important national issue) 2. Lies to customers who paid to use Huawei models 3. A rigid, KPI-focused and unthinking organization where dishonest gaming of the performance review system not only works but seems to be the whole point and is tacitly approved (this and the reporters idealism and loss of faith is the main point of the report as I see it)
- Lol, what? So is this literally a part of CCP military?
https://9to5mac.com/2025/07/04/apple-just-released-a-weirdly... https://news.ycombinator.com/item?id=44472062
HW strategy/culture has been burning tier1 talent since forever. I remember in the 90s When HW and other domestic PRC telco started poaching from Nortel, Siemens, Lucent etc... the talent (most Chinese diaspora used to comfy western office culture) did not have a good time fitting into an actual Chinese company with Chinese culture (but got paid lots). Many burned out too... yet HW, a particularly extreme outlier of militant work culture, has become dominant..
LBH, both HW post sanctions, is a strategic company, overlapping with semi fabrication, domestic chips, and AI is cubing their strategic value. They can get away with doing anything under the current geopolitical environment to stay dominant. The worthwhile take away from this farewell letter is HW threw enough talent at Ascend that it kind of works now, and potentially can throw enough talent at it to be competitive with Nvidia. AKA how it has always operated, like massive wankers. The intuition from the author and most of us is... you need to reward employees right, cultivate proper workplace environment blah blah blah... but look at HW for the past 30 years. They pay a lot of smart people (including patriotic suckers) A LOT of money, throw them at problems until they break. And win.
tengbretson•4h ago
brookst•4h ago
tonyedgecombe•4h ago
npteljes•4h ago
didibus•4h ago
lambdasquirrel•4h ago
didibus•5m ago
No one knows how any of the models got made, their training data is kept secret, we don't know what it contains, and so on. I'm also pretty sure a few of the main models poached each others employees which just reimplemented the same training models with some twists.
Most LLMs are also based on initial research papers where most of the discovery and innovation took place.
And in the very end, it's all trained on data that very few people agreed or intended would be used for this purpose, and for which they all won't see a dime.
So why not wrap and rewrap models and resell them, and let it all compete for who offers the cheapest plan or per-token cost?
esskay•4h ago
some_random•4h ago
perching_aix•4h ago
Arainach•4h ago
mensetmanusman•2h ago
Arainach•1h ago
pton_xd•4h ago
Outright theft is a meaningless term here. The new rules are different.
The AI space is built on "traditionally" bad faith actions. Misappropriation of IP by using pirated content and ignoring source code licenses. Borderline malicious website scraping. Recitation of data without attribution. Copying model code / artifacts / weights is just the next most convenient course of action. And really, who cares? The ethical operating standards of the industry have been established.