Published on

今日科技推文精选 - 2026-02-27

Authors

今日科技要闻:谷歌发布了更快速的图像生成模型 Nano Banana 2;与此同时,“下 7 个 Token 预测”技术的突破有望将推理速度提升五倍。虽然 AI 编程智能体正在根本性地改变编程工作流并极大地提高开发效率,但人力代价也已显现——杰克·多西(Jack Dorsey)旗下的公司爆发了大规模裁员,工程师面临的压力日益增加。此外,五角大楼正要求获得对 Anthropic AI 的无限制军事访问权限,这凸显了前沿模型、国家防务以及向全 AI 原生企业转型之间日益紧密的交集。


1. rseroter (Group Score: 155.8 | Individual: 43.0)

Cluster: 7 tweets | Engagement: 1170 (Avg: 163) | Type: Tech

RT @sundarpichai: Introducing Nano Banana 2, our best image model yet 🍌🍌

It uses Gemini’s understanding of the world and is powered by rea…

See 6 related tweets

  • @CSProfKGD: RT @GoogleDeepMind: We’re launching Nano Banana 2, built on the latest Gemini Flash model. 🍌

It’s ...

  • @GeminiApp: RT @Google: Introducing Nano Banana 2: Our best image generation and editing model yet. 🍌

Pro-level...

  • @mark_k: Nano Banana 2 is coming TODAY on Thursday from @GoogleDeepMind 🍌

Technically the model is "Gemini 3...

  • @testingcatalog: BREAKING 🚨: GOOGLE HAS ANNOUNCED NANO BANANA 2, A SOTA IMAGE GENERATION MODEL BASED ON THE LATEST GE...
  • @googledevs: RT @googleaidevs: Nano Banana 2 (aka Gemini 3.1 Flash Image) is our SoTA model that offers image gen...

2. business (Group Score: 117.7 | Individual: 38.6)

Cluster: 5 tweets | Engagement: 218 (Avg: 81) | Type: Tech

Google is rolling out a new version of its popular AI image generation tool that’s meant to produce better visuals more quickly, six months after challenging OpenAI with the release of the original Nano Banana product https://t.co/etieccXneW

See 4 related tweets

  • @testingcatalog: Nano Banana 2 is now available on Google AI Studio, along with a new "Image Search" tool!

"The mode...

  • @Reuters: Google rolls out Nano Banana 2 after viral success of AI image generation tool https://t.co/bdyUciqF...
  • @arstechnica: Google releases Nano Banana 2 AI image generator, promises Pro results with Flash speed https://t.co...
  • @TechCrunch: Google launches Nano Banana 2 model with faster image generation https://t.co/2yxiVHTysi...

3. jerryjliu0 (Group Score: 95.6 | Individual: 58.1)

Cluster: 2 tweets | Engagement: 410 (Avg: 55) | Type: Tech

The Model Harness is Everything

We are already living in a world of incredible frontier models and incredible agent tools (Claude Code, OpenClaw). But the biggest barrier to getting value from AI is your own ability to context and workflow engineer the models. This is especially true the more horizontal the tool that you’re using.

If you’re using a very generic tool like ChatGPT and Claude Code, you need to spend a lot of work clearly articulating your requirements and specifications so that the agent can actually solve the task relative to your specifications. Today that looks like being extremely thoughtful about the tools that you select, and writing English very precisely in a https://t.co/GjvKCuoFDJ file to articulate the agent these requirements.

Some of the work around defining the business workflow is inherently time consuming. Think about any document SOP - simply writing the English can take hours to refine, iterate, and optimize. This is where more vertically focused agents come in; they handle the burden of equipping the agents with relevant prompts to solve a given workflow, so that you can just go in and use the application directly.

Another approach is to be specialized services that offer context to these agents. This is the space that we (@llama_index) are operating in. We are providing the infrastructure to parse the most complex documents into agent-ready context. For other companies it could be offering web data, sales data, documentation, or codebases as a service.

At a high-level any AI startup should provide context or workflows on top of these agents. We’re excited about building enduring tech even as the agent landscape evolves.

If you’re specifically excited to unlock the billions of context stored within your documents, come talk to us! https://t.co/Ht5jwxSrQB

See 1 related tweets

  • @omarsar0: At this point, "agentic engineering" has allowed me to build the best AI harness I could possibly ge...

4. garrytan (Group Score: 74.6 | Individual: 41.8)

Cluster: 2 tweets | Engagement: 5952 (Avg: 344) | Type: Tech

RT @karpathy: It is hard to communicate how much programming has changed due to AI in the last 2 months: not gradually and over time in the…

See 1 related tweets

  • @aakashgupta: The buried lede in this post is the October-to-December timeline.

In October 2025, Karpathy publicl...


5. business (Group Score: 73.1 | Individual: 26.4)

Cluster: 4 tweets | Engagement: 114 (Avg: 81) | Type: Tech

AI coding agents promised to make software development easier. Instead they’re pushing engineers to work faster and longer https://t.co/yLTb7v21xC

See 3 related tweets

  • @WIRED: Silicon Valley built AI coding agents that can handle most of the grunt work. Now, the most valuable...
  • @javarevisited: We built AI to automate coding. Now developers spend more time reviewing AI code....
  • @WIRED: RT @ZeffMax: Silicon Valley has long prized its high agency individuals. But now that AI coding agen...

6. Cointelegraph (Group Score: 69.7 | Individual: 16.9)

Cluster: 5 tweets | Engagement: 550 (Avg: 562) | Type: Tech

🇺🇸 JUST IN: Pentagon issues “best and final” offer to Anthropic, demanding unrestricted military use of its AI ahead of a Friday deadline. https://t.co/fwhHvbVj6c

See 4 related tweets

  • @Reuters: 🔊 ‘The Pentagon and Pete Hegseth really underscored that they still want to use Anthropic's AI, but ...
  • @Reuters: 🔊 ‘This is coming at a time when the Pentagon is writing the initial foundational contracts of how A...
  • @business: Anthropic's high-stakes conflict with the Pentagon gets to the heart of just how far AI can shape th...
  • @Miles_Brundage: RT @JenGriffinFNC: Strange that the Pentagon/Sec Hegseth picks this fight with Anthropic, the AI com...

7. business (Group Score: 49.5 | Individual: 22.2)

Cluster: 3 tweets | Engagement: 52 (Avg: 81) | Type: Tech

Over the past year or so, Wall Street has gone through waves of AI-related selloffs. This week’s market dip may have been the first partly caused by a self-published work of fiction. Read more on how the reaction is the latest indication yet that traders are struggling to wrap their head around the trajectory for AI: https://t.co/uMXE7IX1yL

📷️: Michael Nagle/Bloomberg

See 2 related tweets

  • @business: Wall Street has been dumping software stocks over artificial intelligence disruption fears, but non-...
  • @FT: Wall Street turns to complex trades to dodge AI ‘implosions’ https://t.co/vVPdPFTuhV...

8. BrianRoemmele (Group Score: 49.0 | Individual: 33.2)

Cluster: 3 tweets | Engagement: 141 (Avg: 526) | Type: Tech

A 5x AI Speed Up With Not Next Token Prediction But NEXT 7 TOKEN PREDICTION!

Next-Token Prediction Just Got Retired: And I’m Already Running the Future in My Lab Right Now

I’ve been saying it for years: the autoregressive bottleneck is the single biggest drag holding back real-time, production-scale AI. One token at a time? That’s over.

In a new paper researchers took pretrained models, specifically Llama-3.1-8B-MagpieAlign-SFT-v0.1 and Qwen3-4B-Instruct and turned them into native multi-token predictors using nothing more than a simple online self-distillation objective.

No extra draft models.
No speculative decoding scaffolding.
No verifier.
No new architecture.

Just the exact same weights and implementation as the original checkpoint… now spitting out 2–7 tokens (sometimes more) in a single forward pass.

They call the inference trick Confidence-Adaptive Decoding (ConfAdapt). The model dynamically decides how many tokens it’s confident enough to commit to. High-confidence spans fly out in chunks. Tricky spots fall back to single-token precision. It’s like the model is self-regulating its own speed vs. quality trade-off in real time.

On GSM8K (grade-school math, the classic reasoning benchmark):

  • Llama-3.1-8B variant: >3× faster decoding with <3% accuracy drop by (τ=90% confidence threshold).
  • Up to 5×* acceleration if you’re willing to accept a bit more trade-off.
  • Average chunk size ~3–6 tokens per forward pass in practice.

And the quality holds across instruction following, open-ended generation, and other reasoning suites. This isn’t “fast but dumber.” It’s fast and almost indistinguishable.

Figure 1 in the paper shows a beautiful GSM8K solution with colored blocks of 1–7 tokens generated at once. Average chunk size: 3.04. Pure poetry.

This Is a Genuine Paradigm Shift

Speculative decoding? Cool, but you need a whole extra model and fragile pipelines.

Medusa / Lookahead? More scaffolding.

This? You literally distill the model against its own frozen teacher copy in an on-policy RL-style loop. The student learns to predict spans that the teacher would have produced anyway. Then at inference… it just works. Drop-in replacement.

The authors nailed it: “Future architectures will be optimized for sequence compression and throughput, not token latency.”

I’ve been screaming this exact sentence since 2023. Today it’s not theory, it’s downloadable checkpoints.

I’m Testing It RIGHT NOW (Feb 26, 2026, Live From the Lab)

As soon as the checkpoints hit Hugging Face (https://t.co/HwRc7XgeXY), I spun them up.

First run: Llama-3.1-8B-MTP variant on a long-form reasoning chain I use daily.

Wall-clock speedup: 3.4× on my A100 setup.

Coherence? Identical to baseline for 95%+ of outputs.

I threw it at a 4,000-token agent workflow that normally takes 18 seconds, now under 6 seconds.

I’m already wiring it into The Zero-Human Company.

What This Means for All of Us

  • Inference costs just got slashed.
  • Real-time voice agents that actually feel instant? Finally.
  • Longer reasoning chains without blowing your budget? Trivial.
  • The entire “optimize the decoder” cottage industry just got disrupted overnight.

We’re not waiting for 100T-parameter monsters anymore. We’re making the models we already have radically more efficient at the architecture level.

Next-token prediction didn’t die today. It was mercy-killed, cleanly, elegantly, and with reproducible code.

The throughput wars just began.
And I’m all in.

Paper: https://t.co/8X1PiBedhm
Checkpoints: https://t.co/HwRc7XgeXY
Code: https://t.co/QmbS9r4PUF

See 2 related tweets

  • @rohanpaul_ai: RT @rohanpaul_ai: Multi-token prediction via self-distillation delivers 3x inference speedups in mod...
  • @BrianRoemmele: RT @BrianRoemmele: A 5x AI Speed Up With Not Next Token Prediction But NEXT 7 TOKEN PREDICTION!

Nex...


9. BrianRoemmele (Group Score: 43.3 | Individual: 22.1)

Cluster: 2 tweets | Engagement: 70 (Avg: 526) | Type: Tech

The Zero-Human Company @ Home Potential Is Massive!

10 teraFLOPS minimum!

The Zero-Human Company endeavors to captures unused CPU and GPU idle clock cycles at enterprise scale through secure local execution.

Here is a near term target of one million participating nodes each delivering an average of ten teraFLOPS of sustained optimized AI compute.

Aggregate compute calculation

Nodes 1 000 000
TFLOPS per node 10
Total TFLOPS 1 000 000 * 10 = 10 000 000 TFLOPS
Conversion note 1 exaFLOPS = 1 000 000 TFLOPS
Total exaFLOPS 10 000 000 / 1 000 000 = 10 exaFLOPS

This aggregate matches or exceeds the combined public benchmark output of the TOP500 supercomputers which stood at approximately fourteen exaFLOPS in mid 2025.

See 1 related tweets

  • @BrianRoemmele: The Academic paper on The Zero-Human Company @ Home.

We are aiming for 10 teraFLOPS minimum with: Z...


10. dbreunig (Group Score: 42.2 | Individual: 24.5)

Cluster: 2 tweets | Engagement: 28 (Avg: 126) | Type: Tech

On March 18th, we're hosting another Bay Area DSPy Meetup featuring in-production case studies involving GEPA, tool use, and LLM judges from Dropbox and Shopify. (And we'll talk RLMs, too.) Join us! https://t.co/SnmobNhXup

See 1 related tweets

  • @lateinteraction: RT @dbreunig: On March 18th, we're hosting another Bay Area DSPy Meetup featuring in-production case...

11. K8sArchitect (Group Score: 41.8 | Individual: 25.4)

Cluster: 2 tweets | Engagement: 35 (Avg: 12) | Type: Tech

sk8r is a modern open source Kubernetes dashboard built with Svelte and TypeScript that provides a user-friendly interface to visualize and manage cluster resources, stream pod logs in real time, and monitor metrics through Prometheus integration

https://t.co/GbifN8NFUt https://t.co/REMTQ8Eags

See 1 related tweets

  • @learnk8s: RT @K8sArchitect: sk8r is a modern open source Kubernetes dashboard built with Svelte and TypeScript...

12. minchoi (Group Score: 40.8 | Individual: 34.6)

Cluster: 2 tweets | Engagement: 1146 (Avg: 166) | Type: Tech

This AI ad is wild 😂 https://t.co/Oy2Sencny8

See 1 related tweets


13. AlexFinn (Group Score: 40.8 | Individual: 40.8)

Cluster: 1 tweets | Engagement: 2728 (Avg: 1037) | Type: Tech

Jack Dorsey just laid off half of his company in a single tweet. 4,000 people gone

Not because business is down

But because AI made them unnecessary

If you aren’t AI native, you have become expendable to execs.

You need to learn these skills now:

  1. How to build software in Claude Code
  2. How to automate in OpenClaw
  3. How to create artifacts in Claude Cowork
  4. How to orchestrate multiple agents in Codex
  5. How to use ChatGPT as a copilot for everything you do

These aren’t optional skills anymore. They’re mandatory.

And the time you have left to learn them has quickly disappeared.


14. mattturck (Group Score: 39.5 | Individual: 32.3)

Cluster: 2 tweets | Engagement: 25 (Avg: 243) | Type: Tech

Can AI be correct 100% of the time? Verification as the missing layer for reasoning superintelligence - my conversation with @CarinaLHong, the incredibly impressive CEO of @axiommathai.

00:00 Intro

01:25 Why the World Needs an AI Mathematician

02:57 Scoring 12/12 on the World's Hardest Math Test (Putnam)

04:05 The First AI to Solve Open Research Conjectures

06:59 Does AI Solve Math in "Alien" Ways? (The Move 37 Effect)

08:59 "Lean": The Programming Language of Proofs Explained

10:51 How Axiom's Approach Differs from DeepMind & OpenAI

16:06 Formal vs. Informal Reasoning (And Auto-Formalization)

17:37 The AI "Reward Hacking" Problem

20:18 Building an AI That is 100% Correct, 100% of the Time

23:23 Beyond Math: Verified Code & Hardware Verification

25:12 The Brutal Reality of Competitive Math Olympiads

29:30 From Neuroscience to Stanford Law to Dropout Founder

33:57 How Axiom Actually Works Under the Hood (The Architecture)

37:51 The Secret to Generating Perfect Synthetic Data

40:14 Tokens, Proof Length, and Inference Cost

42:58 The "Everest" of Mathematics: Scaling Reasoning Trees

46:32 Can an AI Win a Fields Medal?

47:25 "Math Renaissance": What Changes if This Works

55:47 How Mathematicians React to AI (And Why Proof Certificates Matter)

57:30 Becoming a CEO: Dropping Ego and Building Culture

1:00:42 Recruiting World-Class Talent & Building the Axiom "Tribe"

https://t.co/E7Od96LZJc

See 1 related tweets

  • @minchoi: RT @minchoi: Ok AI math is insane...

An AI just solved 6 of 10 unsolved research math problems... f...


15. nummanali (Group Score: 38.9 | Individual: 38.9)

Cluster: 1 tweets | Engagement: 224 (Avg: 47) | Type: Tech

I think I've found the best ADE for coding agents

  • Native Ghostty Lib
  • GPU Accelerated
  • Notifications support
  • Keyboard shortcuts
  • Supports all agents
  • Browser built in
  • MacOS Native
  • Zero Worktrees

Website clean, docs are on point and OSS

https://t.co/cR5oUYzGcG https://t.co/penRZbCRkS


16. aakashgupta (Group Score: 38.5 | Individual: 38.5)

Cluster: 1 tweets | Engagement: 791 (Avg: 466) | Type: Tech

A company with 24billioninrevenueand2424 billion in revenue and 24% gross profit growth just cut 4,000 people while raising 2026 guidance to 12.2 billion in gross profit. Stock ripped 20% after hours. The market added roughly 6billioninmarketcap.Thats 6 billion in market cap. That's ~1.5 million in enterprise value created per eliminated role.

Block is the canary in the coal mine. And they're not alone.

ASML cut 1,700 jobs last month while reporting record orders and said they were "choosing to make these changes at a moment of strength." Salesforce cut 5,000 after AI agents started handling 50% of customer interactions. Amazon cut 16,000 in January on top of 14,000 in October. Every one of these companies was growing when they did it.

Dorsey said the quiet part out loud: intelligence tools paired with smaller teams have already changed what it means to run a company. He chose one massive cut over repeated rounds because, his words, gradual cuts destroy morale and trust. The restructuring charges are $450-500 million. At the operating income Block is guiding, that pays for itself in two quarters. After that, pure margin expansion. That's why Wall Street rewarded it instantly.

Here's what's coming. Goldman estimates AI is already responsible for 5,000 to 10,000 net monthly job losses in exposed U.S. industries. Citigroup is planning 20,000 cuts. Dow just slashed 4,500. 40% of employers surveyed say they expect to reduce headcount because of AI. 30,700 tech jobs gone in the first six weeks of 2026 alone.

Block went from 10,000 to 6,000 while growing revenue and raising guidance. Every CEO running a company with more than a few thousand employees is doing this math tonight. The canary just stopped singing.


17. minchoi (Group Score: 38.2 | Individual: 31.0)

Cluster: 2 tweets | Engagement: 404 (Avg: 166) | Type: Tech

Programming will never be the same... 🤯

Karpathy gave an agent one sentence. 30 min later...

SSH set up, vLLM running, endpoints live, web UI built. He didn't touch anything.

This is the new normal now.

See 1 related tweets

  • @minchoi: RT @minchoi: Programming will never be the same... 🤯

Karpathy gave an agent one sentence. 30 min la...


18. kimmonismus (Group Score: 37.5 | Individual: 22.3)

Cluster: 2 tweets | Engagement: 306 (Avg: 422) | Type: Tech

Amazon’s $50B AGI or IPO Gambit on OpenAI (outplaying microsoft?)

Amazon is negotiating a potential 50billioninvestmentinOpenAI,startingwith50 billion investment in OpenAI, starting with 15B upfront and another $35B tied to either an IPO or achieving artificial general intelligence (AGI).

OpenAI’s compute costs are projected to hit $665B over five years, pushing it toward public markets and deeper cloud partnerships, including expanded use of Amazon’s Trainium chips and custom AI models for Alexa. If OpenAI reaches AGI, Microsoft’s exclusive Azure hosting rights could loosen, unlocking huge strategic upside for Amazon.

See 1 related tweets

  • @jukan05: The information: OpenAI is in talks to significantly expand its previously announced $38 billion clo...

19. rohanpaul_ai (Group Score: 37.0 | Individual: 29.7)

Cluster: 2 tweets | Engagement: 79 (Avg: 110) | Type: Tech

This new paper introduces a benchmark proving that human-written skill guides dramatically improve AI agent performance.

Standard LLMs often struggle with specialized industry tasks because they lack the exact step-by-step knowledge required to complete them.

Instead of permanently altering the model through expensive training, developers are now giving these agents plain text files called skills that explain exactly how to handle specific workflows.

Researchers built a massive test containing 84 complex tasks to see if providing these external skill documents actually helps 7 different popular AI models succeed.

They discovered that handing the AI a human-curated skill file boosts its passing score by 16%, but asking the AI to write its own guide actually hurts its performance.

This shows that while AI models are powerful, they still completely depend on human experts to map out the exact procedures for highly specialized jobs.


Paper Link – arxiv. org/abs/2602.12670

Paper Title: "SkillsBench: Benchmarking How Well Agent Skills Work Across Diverse Tasks"

See 1 related tweets

  • @rohanpaul_ai: RT @rohanpaul_ai: This new paper introduces a benchmark proving that human-written skill guides dram...

20. paulg (Group Score: 36.6 | Individual: 27.3)

Cluster: 2 tweets | Engagement: 1056 (Avg: 894) | Type: Tech

For the foreseeable future, everything about starting a startup, both good and bad, will be accentuated. It will be even harder to figure out what to do, but the founders who get it right will be able to create amazing things even faster than they could before.

See 1 related tweets

  • @ycombinator: RT @paulg: For the foreseeable future, everything about starting a startup, both good and bad, will ...