*edit: not that it matters, but since MAGA can't help but assume, these are all US citizens and green card holders that I am referring to.
https://apnews.com/article/immigration-raid-hyundai-korea-ic...
https://www.koreatimes.co.kr/foreignaffairs/20251112/hundred...
https://www.pbs.org/newshour/nation/attorney-says-detained-k...
The regime is powered by racism and doesn't think through things.
The risk and level of publicity is just too high for many people to even consider, especially people already intelligent/capable enough to immigrate anywhere else that doesn't have these issues or stay in their own country.
1. Detained the incorrect person
2. Detained the correct person, with the correct legal status
3. Detained the correct person, with the correct legal status, but in unlawful circumstances
4. Detained the correct person, with the correct legal status, in ostensibly-lawful circumstances, but in a way which is unconstitutional or crazy
An example of the final category are the immigrants that spent years being vetted, following the law, and doing expensive paperwork to be citizens. ICE snatched them when they showed up on at the last second as they were to take their citizenship oath. [0] Not because of anything they did, but because today's Republican party has decided that it's OK to hurt people based on their "shithole" country of birth.
[0] https://www.theguardian.com/us-news/2025/dec/30/us-citizensh...
If anything I find the stories featuring white/European people oddly racist because they seem to assume that I, the reader, will assume a white/European person couldn't possibly be in violation of immigration rules. But all the ones I've read turned out that they were indeed in violation of immigration rules.
Overall as a potential immigrant to the US myself, I find the process capricious and that US citizens by birth don't fully appreciate how painful it is or why it shouldn't be that way. But I don't find it notably worse or more onerous than the vast majority of immigration policies of other countries in practice.
[1] https://www.theguardian.com/us-news/2026/jan/13/ice-immigrat...
[2] https://www.pbs.org/newshour/nation/a-u-s-citizen-says-ice-f...
[3] https://www.propublica.org/article/immigration-dhs-american-...
To address your stories specifically, my point would be that I'm still not sure whether this shows the US is notably worse on this than any other place.
Their goal is for every one person violently detained, 10 decide to leave on their own, and 100 decide to not come in the first place.
If they actually thoroughly evicted non-status migrant workers they'd have a outright revolt on their hands from farmers and other businesses that depend on them.
Instead those businesses can now take further advantage of the fear of harassment and/or deportation to drive down compensation and rights.
Contrast with countries like Canada that have a legal temporary foreign agriculture worker program that provides a regulated source of seasonal migrant farm worker labour under a non-citizen temporary status, but with some rights (still often abused). It's notable to me as a Canadian that I don't see this being advocated on any large scale by either party in the US.
Anyways, all this just to say that the jackboot clown theater is the point, not a side effect.
Because going around and harassing and deporting other or non-essential non-status immigrants would drive labor costs down because of the chill it would put through those who are grudgingly tolerated.
And besides, given the quality of personality ICE seems to be employing even (especially) at its highest levels, I simply assume there's corruption such that if I'm a large orchard or whatever I simply pay ICE to stay away.
"There was a significant drop-off in entries to the United States in 2025 relative to 2024 and an increase in enforcement activity leading to removals and voluntary departures. We estimate that net migration was between –10,000 and –295,000 in 2025, the first time in at least half a century it has been negative."
There'd be no food on the tables, frankly. And people in Silicon Valley would have messy houses and algae in their pools.
Soybean farmers are screwed.
I dont think youre as right as you want to believe. Certainly not as right as I want you to believe
In any case, there's no way Anthropic's investors in Silicon Valley would countenance such a move.
Also, I'm biased the logical place is Canada, not Europe. Much of the fundamental/foundational research on LLMs, and a large part of the talent, came from universities in Canada anyways.
Europe really just needs to rally behind Mistral. That's where they should dump their cash.
At the moment my impression is instead that the issue is computational resources. It's important to stay near the frontier though, and to build up ones capacity to train large models.
Consequently I don't think we need Anthropic. It wouldn't be terrible if they came. Especially if they picked a nice location. Barcelona would be very nice, for example.
Besides you can live a comfortable life in PRC nowadays or live in a racist America.
When I was a deep learning PhD in the first Trump administration, US universities were already very deeply affected by the Muslim ban, and so a lot of talent ended up in other countries.
Sibling commentators are rightfully pointing out that foreigners, especially those who would not be recognized as white, face an onerous and risky customs process with long-term and increasing risks of deportation. When you see a headline like the NIST labs abruptly restricting foreign scientists, _everything_ else feels uncertain. Even if someone doesn't believe they're personally at risk for deportation, they're still seeing everything else.
And then it all boils down to a reputational thing. The era where we were the top choice for research is in the past. If you start a PhD in the US on your resume during this era, you might be anticipating how you'll answe the question of why you weren't good enough to get accepted somewhere better.
I will say we are winning in accessibility. China doesn’t have much of a ramp game
I wonder if you max out your options in China. It seems the Party is suspicious of ambition and high profile winners. I'm sure you can live comfortably, but there's a ceiling.
Isn't it just straight-up illegal in China to refuse the government from using your model? USA isn't perfect, but at least it has active discourse.
People in Hong Kong died. Over 10,000 were arrested and many are still in prison. The rest are permanently disgraced in their social-credit society.
Again, USA is not perfect, but let's not dream up some fantasy about the CCP.
https://reclaimthenet.org/china-man-chair-interrogation-soci...
Does that matter? In China people don't judge the state of their civilization by how easily you can insult the police but whether you need to be afraid to meet them on the street. "I can insult my pedophile president" (who doesn't care if you do) isn't exactly a flex.
It does tell us something though that the evaluation of American life now consists of parasocial interactions with the president on social media. I'm starting to belief Bruno Maçães, ex Portuguese secretary of state, was prescient with his diagnosis that American material society has rotted to the point where life is now entirely defined by virtual interactions. That's the difference between China and the US today.
The president's a pedophile, a criminal, undeterred by democracy, economy or social disorder but you can freely yell into the void. Have you considered that in the US one can freely say all these things precisely because that's irrelevant?
Americans will vote for their Congress representatives in November. They will have a chance to decide how they want their government to be run. The US President was already shot-down once by the Supreme Court (tariffs). The system is working. Let the voters decide, and then let it work.
But this:
> According to the social credit system, Chinese citizens are punishable if they indulge in buying too many video games, buying too much junk food, having a friend online who has a low credit score, visiting unauthorized websites, posting “fake news” online, and more.
...is just pure bullshit. There were _ideas_ about including these kinds of stuff into the score, but they have never been implemented. At this point, the social credit score is only used to find people who dodge court decisions.
Please ignore the gun pointed at your head / social credit score / masked goons roving about Minnesota / flock cameras / etc as it hasn't been used against you at this point.
Do you imagine an invasion of Taiwan won't involve dropping bombs?
I feel like we should be able to agree that providing authoritarian regimes with high tech tools is immoral in the general case.
If you'd asked me two years ago my answer might have been different.
And to the original point, yeah, I would feel entirely justified in the critique of engineers in providing tools to the US defense apparatus at this point.
At least the Chinese shops are giving their weights away for free, and not demanding that any government ban the rest.
In the US people try to hide it and are far more sinister about it, since there are a lot of laws against obvious racism. The cops are also happy in the US to just kill you.
The racism in the US comes out of hate where as what I experienced abroad was more, we don't think you'll fit in and follow the rules and you have to constantly prove that you can.
I didn't spend too much time in China so maybe it is a racist hell hole.
But my experience in Japan was that white immigrants were way more inclined to make a huge deal about the lighter racism they experienced because they had never been somewhere where their skin color was a disadvantage.
I speculate that if you were a permanent minority instead of a visiting inconvenience, then that 'nice' racism you describe would metastasize into the type of racism you see in the USA. It's more friction from time and exposure added on. And, you know, slavery.
I'm sure it's a very nice place to live if you're content to just stay quiet in society and never put a political sign in your yard or even just talk about the wrong thing with your friend in a WeChat.
In practical terms, if you're not kind of person who would want to run for an office in the US, China is incredibly comfortable. Cities are safe, with barely any violent crime. Public drug use is nonexistent. And with the US-level AI researcher income, you'd be in the top 0.1% earners.
https://news.ycombinator.com/item?id=47252833
My comment and the linked video says otherwise. The guy was in a private group chat and said some nasty things about the police for confiscating his motorcycle. Now he's arrested and in the Tiger Chair.
How are we explaining this?
Not as bad as China sure, but not as good as other civilized nations.
If you want to put this to the test try crossing the Canadian border and when they ask you the purpose of your visit respond that it's to attend a protest.
Well duh, as recently demonstrated, an US model used by the US gov will 100% end up murdering actual children sooner than later, in this case less than a calendar year in some far flung war that many Americans do not support. Alternatively PRC model used by CCP might kill in some hypothetical future but for national reunification/rejuvenation that many Chinese support. At the end of the day, researchers and population on one side sleeps more soundly.
I think Alibaba needs to just give these guys a blank check. Let them fill it in themselves. Absent that, I'm pretty sure they'll make their own startup.
I do think it'd be a big loss for the rest of the world though if they close whatever model their startup comes up with.
That's very likely to happen once the gap with OpenAI/Anthropic has been closed and they managed to pop the bubble.
If models like Qwen can get good enough for coding tasks locally, the real shift might be economic rather than purely capability.
To be honest, it's sort of what I expected governments to be funding right now, but I suppose Chinese companies are a close second.
Wild times!
If AI could effectively replace people, you wouldn’t need CEOs to keep trying to convince people.
Probably good to sent alerts early, but they might be going a bit too early.
Is there a better agentic coding harness people are using for these models? Based on my experience I can definitely believe the claims that these models are overfit to Evals and not broadly capable.
They also struggle at translating very broad requirements to a set of steps that I find acceptable. Planning helps a lot.
Regarding the harness, I have no idea how much they differ but I seem to have more luck with https://pi.dev than OpenCode. I think the minimalism of Pi meshes better with the limited capabilities of open models.
If so, I'm happy that the team held together, and I hope that endogenous tech leads get to control their own career and tech destiny after hard work leads to great products. (It's almost as inspiring as tank man, and the tank commanders who tried to avoid harming him...)
(ducking the downvote for challenging the primacy of equity...)
But I'll be running this locally for note summarization, code review, and OCR. Very coherent for its size.
the qwen is dead, long live the qwen.
I am trying super hard to use cheap models, and outside SOTA models, they have been more trouble than they are worth.
I really recommend trying the Qwen models - 3 coder next is really incredible. GLM 4.7 flash is also incredibly performant on modest hardware. Important things to consider is setting the temperature and top_p and top_k values etc based on what is recommended by the provider of the model - a thing as simple as that could result in a huge difference in performance.
The other big leap for me was switching to Zed editor and getting its agent stuff just seamlessly integrated. If you run LM Studio on your local machine it's super easy and even setting it up on a remote machine and calling out to LM Studio is dead simple.
Use case means everything. I doubt this model would fare well on a large codebase, but this thing is incredible.
Maybe Qwen3.5-35B-A3B is that model? This comment reports good results: https://news.ycombinator.com/item?id=47249343#47249782
I need to put that through its paces.
So far none of them have be useful enough at first glance with a local model for me to stick with them and dig in further.
It has been useful for education ("What does this Elixir code do? <Paste file> ..... <general explanation> "then What this line mean?")
as well as getting a few basic tests written when I'm unfamiliar with the syntax. ("In Elixir Phoenix, given <subject under test, paste entire module file> and <test helper module, paste entire file> and <existing tests, pasted in, used both for context and as examples> , what is one additional test you would write?")
This is useful in that I get a single test I can review, run, paste in, and I'm not using any quota. Generally I have to fix it, but that's just a matter of reading the actual test and throwing the test failure output to the LLM to propose a fix. Some human judgement is required but once I got going adding a test took 10 minutes despite being relatively unfamiliar with Elixir Phoenix .
It's a nice loop, I'm in the loop, and I'm learning Elixir and contributing a useful feature that has tests.
I've been testing Qwen3.5-35B-A3B over the past couple of days and it's a very impressive model. It's the most capable agentic coding model I've tested at that size by far. I've had it writing Rust and Elixir via the Pi harness and found that it's very capable of handling well defined tasks with minimal steering from me. I tell it to write tests and it writes sane ones ensuring they pass without cheating. It handles the loop of responding to test and compiler errors while pushing towards its goal very well.
The main quirk I've found is that it has a tendency to decide halfway through following my detailed instructions that it would be "simpler" to just... not do what I asked, and I find it has stripped all the preliminary support infrastructure for the new feature out of the code.
That's likely coming from the 3:1 ratio of linear to quadratic attention usage. The latest DeepSeek also suffers from it which the original R1 never exhibited.
> Blah blah blah (second guesses its own reasoning half a dozen times then goes). Actually, it would be a simpler to just ...
Specifically on Antigravity, I've noticed it doing that trying to "save time" to stay within some artificial deadline.
It might have something to do with the system messages and the reinforcement/realignment messages that are interwoven into the context (but never displayed to end-users) to keep the agents on task.
That sounds too close to what I feel on some days xD
That would seem logical, as the results are then completely deterministic, but it turns out that a suboptimal token may result in a better answer in the long run. Also, allowing for a little bit of noise gives the model room to talk itself out of a suboptimal path.
I wonder if determinism will be less harmful to diffusion models because they perform multiple iterations over the response rather than having only a single shot at each position that lacks lookahead. I'm looking forward to finding out and have been playing with a diffusion model locally for a few days.
For creative things or exploratory reasoning, a temperature of 0.8 lends us to all sorts of excursions down the rabbit hole. However, when coding and needing something precise, a temperature of 0.2 is what I use. If I don’t like the output, I’ll rephrase or add context.
This is my experience with the Qwen3-Next and Qwen3.5 models, too.
I can prompt with strict instructions saying "** DO NOT..." and it follows them for a few iterations. Then it has a realization that it would be simpler to just do the thing I told it not to do, which leads it to the dead end I was trying to avoid.
It's also driving itself crazy with deadpool & deadpool-r2d2 that it chose during planning phase.
That said, it does seem to be doing a very good job in general, the code it has created is mostly sane other than this fuss over the database layer, which I suspect I'll have to intervene on. It's certainly doing a better job than other models I'm able to self-host so far.
I think this is part of the model’s success. It’s cheap enough that we’re all willing to let it run for extremely long times. It takes advantage of that by being tenacious. In my experience it will just keep trying things relentlessly until eventually something works.
The downside is that it’s more likely to arrive at a solution that solves the problem I asked but does it in a terribly hacky way. It reminds me of some of the junior devs I’ve worked with who trial and error their way into tests passing.
I frequently have to reset it and start it over with extra guidance. It’s not going to be touching any of my serious projects for these reasons but it’s fun to play with on the side.
I can live with this on my own hardware. Where Opus4.6 has developed this tendency to where it will happily chew through the entire 5-hour allowance on the first instruction going in endless circles. I’ve stopped using it for anything except the extreme planning now.
I’m trying to use local models whenever possible. Still need to lean on the frontier models sometimes.
> Do you feel you could replace the frontier models with it for everyday coding? Would/will you?
Probably not yet, but it's really good at composing shell commands. For scripting or one-liner generation, the A3B is really good. The web development skills are markedly better than Qwen's prior models in this parameter range, too.
I'm aligning on Agent for the combination of harness + model + context history (so after you fork an agent you now have two distinct agents)
And orchestrator means the system to run multiple agents together.
This terminology is still very much undefined though, so my version may not be the winning definition.
It's really easy to setup with any OpenAI compatible API and I self host Qwen Coder 3 Next on my personal MBP using LM Studio and just dial in from my work laptop with Zed and tailscale so i can connect from wherever i might be. It's able to do all sorts of things like run linting checks and tests and look for issues and refactor code and create files and things like this. I'm definitely still learning, but it's a pretty exciting jump from just talking to a chat bot and copying and pasting things manually.
[0] https://www.reddit.com/r/LocalLLaMA/comments/1rivckt/visuali...
Qwen3.5-35B-A3B means that the model itself consists of 35 billion floating point numbers - very roughly 35GB of data - which are all loaded into memory at once.
But... on any given pass through the model weights only 3 billion of those parameters are "active" aka have matrix arithmetic applied against them.
This speeds up inference considerably because the computer has to do less operations for each token that is processed. It still needs the full amount of memory though as the 3B active it uses are likely different on every iteration.