H Company Holo2: Achieves #1 in UI Localization Benchmark

235B Parameter Model Revolutionizes UI Automation

  • Achieves SOTA with 78.5% on ScreenSpot-Pro benchmark
  • Agent localization improves performance by 10-20%
  • Accurately locates small UI elements even on 4K high-resolution interfaces

What Happened?

H Company released Holo2-235B-A22B, a specialized model for UI Localization (identifying the position of user interface elements). [Hugging Face] This 235B parameter model finds the exact location of UI elements like buttons, text fields, and links in screenshots.

The key is Agentic Localization technology. Instead of providing all answers at once, it improves predictions across multiple steps. This allows it to accurately identify even small UI elements on 4K high-resolution screens. [Hugging Face]

Why Is It Important?

The GUI agent field is heating up. Big tech companies like Claude Computer Use and OpenAI Operator are competing to release UI automation features. However, H Company, a small startup, has taken the top spot in this benchmark.

What I personally find noteworthy is the agentic approach. Traditional models often failed when trying to adjust positions in one go, but the approach of improving the model through multiple attempts proved effective. The 10-20% performance improvement demonstrates this.

Frankly, 235B parameters is quite heavy. How fast it runs in actual production environments remains to be seen.

What Will Happen in the Future?

As GUI agent competition intensifies, UI Localization Accuracy is expected to become a key differentiating factor. Since the H Company model was released as open source, it is likely to be integrated into other agent frameworks.

It could also impact the RPA (robotic process automation) market. Traditional RPA tools were rule-based, but now vision-based UI understanding could become the standard.

Frequently Asked Questions (FAQ)

Q: What exactly is UI Localization?

A: It is a technology that looks at a screenshot and finds the exact coordinates of specific UI elements (buttons, input fields, etc.). Simply put, it is AI knowing where to click when looking at a screen. This is a core technology for GUI automation agents.

Q: What is different from existing models?

A: Agentic localization is the key. Instead of trying to get it right in one go, it refines over multiple steps. Similar to how humans scan a screen to find a target. This method achieved a 10-20% performance improvement.

Q: Can I use the model directly?

A: It is publicly available for research on Hugging Face. However, as a 235B parameter model, it requires significant GPU resources. It is more suitable for research or benchmarking purposes rather than actual production applications.


If you found this article useful, please subscribe to AI Digester.

References

Lotus Health AI Raises $35 Million for Free AI Doctor

Free AI Primary Care Doctor Raises $35 Million

  • Lotus Health AI raises $35 million in Series A from CRV and Kleiner Perkins
  • Offers free 24/7 primary care service in 50 languages across all 50 US states
  • In an era where 230 million people ask ChatGPT health questions weekly, serious competition in the AI healthcare market begins

What Happened?

Lotus Health AI raised $35 million in a Series A round co-led by CRV and Kleiner Perkins.[TechCrunch] This startup uses Large Language Models (LLMs) to provide free 24/7 primary care services in 50 languages.

Founder KJ Dhaliwal sold South Asian dating app Dil Mil for $50 million in 2019.[Crunchbase] He was inspired by his childhood experience translating medical information for his parents. Lotus Health AI launched in May 2024 to address inefficiencies in the US healthcare system.

Why Is It Important?

Honestly, this investment amount is notable. The average funding for AI healthcare startups is $34.4 million, and Lotus Health AI reached this level at Series A.[Crunchbase]

The context makes it understandable. According to OpenAI, 230 million people ask ChatGPT health-related questions every week.[TechCrunch] This means people are already getting health advice from AI. But ChatGPT cannot provide medical services. Lotus Health AI targets this gap market.

Personally, the “free” model is most interesting. Considering how expensive US healthcare is, free primary care is a quite disruptive value proposition. Of course, the revenue model is still unclear.

What Will Happen in the Future?

Competition in the AI healthcare market is expected to intensify. OpenAI also entered this market in January with the launch of ChatGPT Health. It integrates with Apple Health, MyFitnessPal, and others to provide personalized health advice.[OpenAI]

Regulatory risks remain. Even OpenAI states in its terms of service “do not use for diagnosis or treatment purposes.” Several lawsuits regarding harm from AI medical advice are already underway. We need to watch how Lotus Health AI manages this risk.

Frequently Asked Questions

Q: Is Lotus Health AI really free?

A: It is free for patients. However, the specific revenue model has not been disclosed. There are various possibilities including B2B models targeting insurance companies or employers, or premium service additions. Since they provide service across all 50 US states, they appear to be pursuing economies of scale.

Q: How is it different from general AI chatbots?

A: Lotus Health AI is a medical service specialized in primary care. Unlike general chatbots, it holds medical service licenses in all 50 US states. The key difference is that it can perform actual medical acts, not just provide health information.

Q: Does it support Korean?

A: It was announced to support 50 languages, but the specific language list has not been disclosed. Korean support needs to be confirmed. Currently, the service is only available in the US, and overseas expansion plans have not been announced.


If this article was helpful, please subscribe to AI Digester.

References

AWS SageMaker Data Agent: Medical Data Analysis, From Weeks to Days

Medical Data Analysis, From Weeks to Days

  • AWS SageMaker Data Agent: An AI agent that analyzes medical data using natural language
  • Perform cohort comparison and survival analysis without writing code
  • Launched November 2025, available for free in SageMaker Unified Studio

What Happened?

AWS unveiled SageMaker Data Agent, an AI agent for medical data analysis. When epidemiologists or clinical researchers ask questions in natural language, the AI automatically generates and executes SQL and Python code.[AWS]

Previously, analyzing medical data required visiting multiple systems to access data. Waiting for permissions, understanding schemas, and writing code manually. This process took weeks. SageMaker Data Agent shortens this to days or hours.

Why Is This Important?

Frankly speaking, medical data analysis has always been a bottleneck. Epidemiologists spent 80% of their time on data preparation and only 20% on actual analysis. The reality was that only 2-3 studies could be conducted per quarter.

SageMaker Data Agent flips this ratio. By significantly reducing data preparation time, researchers can focus on actual clinical analysis. I believe this will directly impact the speed of discovering patient treatment patterns.

What is particularly impressive is that complex tasks like cohort comparison and Kaplan-Meier survival analysis can be requested in natural language. Say “Analyze the survival rates of male viral sinusitis patients versus female patients,” and the AI automatically plans, writes code, and executes it.[AWS]

How Does It Work?

SageMaker Data Agent operates in two modes. First, you can generate code with inline prompts directly in notebook cells. Second, the Data Agent panel breaks down complex analysis tasks into structured steps and processes them.[AWS]

The Agent understands the current notebook state, data catalog, and business metadata, then generates context-appropriate code. Rather than spitting out code fragments, it establishes a complete analysis plan.[AWS]

What Lies Ahead?

According to a Deloitte survey, 92% of healthcare executives are investing in or experimenting with generative AI.[AWS] Demand for medical AI analysis tools will continue to grow.

Agentic AI like SageMaker Data Agent could positively impact drug development and treatment pattern discovery by accelerating medical research. However, one concern is data quality. No matter how fast AI is, garbage in means garbage out.

Frequently Asked Questions (FAQ)

Q: How much does SageMaker Data Agent cost?

A: SageMaker Unified Studio itself is free. However, actual computing resources (EMR, Athena, Redshift, etc.) are charged based on usage. Notebooks have a free tier of 250 hours for the first two months, so you can test it lightly.

Q: What data sources are supported?

A: It connects to AWS Glue Data Catalog, Amazon S3, Amazon Redshift, and various data sources. If you have existing AWS data infrastructure, you can connect immediately. It is also compatible with medical data standards FHIR and OMOP CDM.

Q: In which regions is it available?

A: It is available in all AWS regions where SageMaker Unified Studio is supported. It is best to check the official AWS documentation for whether the Seoul region is supported.


If you found this article useful, please subscribe to AI Digester.

References

I Shredded Millions of Books to Build Claude: The Truth Behind Anthropic Project Panama

$1.5 Billion Settlement, Millions Destroyed: Key Takeaways

  • Anthropic purchased, disassembled, scanned, and destroyed millions of books for Claude training
  • Internal document: “Project Panama is our effort to destructively scan books from around the world”
  • $1.5 billion settlement, approximately $3,000 per book paid to authors

What Happened?

Over 4,000 pages of court documents were released, revealing Anthropic’s secret project. The codename was “Project Panama.” Internal planning documents explicitly state: “Project Panama is our effort to destructively scan books from around the world.” They bulk-purchased tens of thousands of books from used bookstores like Better World Books and World of Books. They cleanly cut off the spines using “hydraulic cutters.” They scanned the pages with high-speed, high-quality scanners. And recycling companies collected the remaining debris.[Techmeme]

This project was led by Tom Turvey. He’s a former Google executive who created the Google Books project 20 years ago. Over about a year, Anthropic invested millions of dollars to acquire and scan millions of books.[Futurism]

Why Is This Important?

Honestly, this reveals the reality of how AI training data is acquired.

Why did Anthropic choose this approach? First, to avoid the risks of illegal downloads. Second, buying used books and disposing of them as desired was likely legal under the “first sale doctrine.” The judge actually recognized this scanning method itself as fair use.[CNBC]

However, there was a problem. Before Project Panama, Anthropic had freely downloaded over 7 million books from illegal sites like Library Genesis and Pirate Library Mirror. The judge ruled that this portion could constitute copyright infringement.[NPR]

Personally, I think this is the core issue. The problem isn’t scanning “legitimately” purchased books to destroy them—it’s that they first illegally downloaded them. Anthropic itself was aware of this. Internal documents explicitly state: “We don’t want this work to be known.” Will there be consequences?

The $1.5 billion settlement is the largest in AI copyright dispute history. Approximately $3,000 per book goes to authors for about 500,000 books.[PBS]

AI has set other precedents. The impact on companies is significant. OpenAI, Google, and Meta also face similar lawsuits. The standard has become clear: “Buying books and scanning them is fine, but illegal downloads are not permitted.”

Anthropic is already embroiled in a music copyright lawsuit. A separate lawsuit was filed in January, with music publishers claiming that Claude 4.5 was trained to “memorize” copyrighted works. Watchdog]

Frequently Asked Questions

Q: How many books were actually usable in Project Panama? Were they destroyed?

A: According to court documents, up to 2 million books were targeted for “destructive scanning.” Anthropic purchased tens of thousands of books from used bookstores like Better World Books and World of Books, and it’s estimated they processed millions of books over about a year, investing millions of dollars.

Q: How much will authors receive?

A: The $1.5 billion settlement applies to approximately 500,000 books. That’s about $3,000 per book. Authors of illegally downloaded books are eligible to file claims, and once the settlement is approved by the court, they can file individually. However, if not all authors file claims, the actual amount received may increase.

Q: Is it legal to buy books and scan them?

A: The judge recognized this method as fair use. This is because, under the “first sale doctrine,” purchased books can be disposed of as desired. However, Anthropic’s problem was that they downloaded books from illegal sites before Project Panama. Scanning legally purchased books is currently legal.


If you found this article useful, subscribe to AI Digester.

References

One Year of DeepSeek: 113,000 Qwen Derivatives, 4x More Than Llama

One Year of Deep Chic Moment: 3 Changes Proven by Numbers

  • Over 113,000 Qwen derivative models — 4x more than Meta Llama (27,000)
  • DeepSeek ranks #1 in Hugging Face followers, Qwen at #4
  • Chinese AI organizations shift direction: “Open source is strategy”

What Happened?

Hugging Face released their ‘Deep Chic Moment’ one-year analysis report.[Hugging Face] This is the final part of a three-part series summarizing data on how China’s open source AI ecosystem has grown since DeepSeek’s emergence in January 2025.

Let’s start with the key metrics. Qwen (Alibaba) derivative models exceeded 113,000 as of mid-2025. Including repositories tagged with Qwen, the number surpasses 200,000.[Hugging Face] This is an overwhelming figure compared to Meta’s Llama (27,000) or DeepSeek (6,000).

Why Is It Important?

Frankly speaking, just a year ago, many people regarded Chinese AI as ‘copycat.’. But now it’s different.

ByteDance, Deepseek, Tencent, and Qwen rank among the top in Hugging Face’s popular papers rankings. In terms of follower count, DeepSeek is #1 and Qwen is #4. Looking at Alibaba as a whole, the number of derivative models matches Google and Meta combined.[Hugging Face]

What I personally find notable is Alibaba’s strategy. Qwen is structured as a ‘family,’ not a single flagship model. It supports various sizes, tasks, and modalities. Simply put, it means: “Use our models as general-purpose AI infrastructure.”

What Will Happen Next?

Hugging Face analyzed that “open source is a short-term dominance strategy for Chinese AI organizations.” The interpretation is that they aim for large-scale integration and deployment by sharing not only models but also papers and deployment infrastructure.

Within just one year, the numbers confirmed that the DeepSeek moment was not a one-time event. The center of gravity in the global AI open source ecosystem is shifting.

Frequently Asked Questions (FAQ)

Q: Are there more Qwen derivatives than Llama? Why?

A: Alibaba released Qwen in various sizes and modalities, expanding its application range. Chinese developers frequently use it for local deployment. The strategy of continuously updating the model range with Hugging Face has also been effective.

Q: Is DeepSeek still important?

A: Yes. DeepSeek has the most followers on Hugging Face. However, it trails Qwen in derivative model count. DeepSeek has strengths in papers and research contributions, while Qwen focuses on ecosystem expansion.

Q: What does this mean for developers?

A: Qwen-based models are strengthening multilingual support. Because it’s open source, local deployment and fine-tuning are free. It’s become a great environment to experiment without cost burden. However, license terms vary by model, so check before use.


If this article was useful, subscribe to AI Digester.

References

OpenAI Reveals Sora Feed Philosophy: “Doomscrolling Is Not Allowed”

OpenAI, Sora feed philosophy revealed: “We do not allow doomscrolling”

  • Creation first, consumption minimization is the key principle
  • A new type of recommendation system that can be adjusted with natural language
  • Safety measures from the creation stage, opposite strategy to TikTok

What happened?

OpenAI officially announced the design philosophy behind Sora’s recommendation feed, their AI video creation app.[OpenAI] The core message is clear: “This is a platform for creation, not doomscrolling.”

While TikTok has faced controversy for optimizing watch time, OpenAI chose the opposite direction. Instead of maximizing feed dwell time, they prioritize showing content most likely to inspire users to create their own videos.[TechCrunch]

Why is it important?

Honestly, this is quite an important experiment in social media history. Existing social platforms maximize dwell time to generate ad revenue. The longer users stay, the more money they make. This has resulted in addictive algorithms and mental health issues.

OpenAI already generates revenue through subscription models (ChatGPT Plus). Since they don’t rely on ads, they don’t need to “keep users hooked.” Simply put, because the business model is different, the feed design can be different too.

Personally, I’m curious whether this will actually work. Can a feed that “encourages creation” really keep users engaged? Or will it eventually revert to dwell time optimization?

4 Principles of Sora Feed

  • Creative Optimization: Induces participation, not consumption. The goal is active creation, not passive scrolling.[Digital Watch]
  • User control: The algorithm can be adjusted with natural language. Commands like “Show me only comedy today” are possible.
  • Connection priority: Content from followers and people you know is shown before viral global content.
  • Safety-freedom balance: Since all content is generated within Sora, harmful content is blocked at the creation stage.

How is it different technically?

OpenAI differs from existing LLMs. Using this approach, a new type of recommendation algorithm was developed. The key differentiator is “natural language instructions.” Users can explain to the algorithm directly in words what type of content they want.[TechCrunch]

Sora uses activity (likes, comments, remixes), IP-based location, ChatGPT usage history (can be turned off), and creator follower count as personalization signals. However, safety signals are also included to suppress harmful content exposure.

What will happen in the future?

The Sora app launched in just 48 hours. It reached #1 on the App Store. 56,000 downloads on the first day, tripled on the second day.[TechCrunch] Initial response was enthusiastic.

But the question is sustainability. As OpenAI acknowledges, this feed is a “living system.” It will continue to change based on user feedback. What happens when the creation philosophy conflicts with actual user behavior? We’ll have to watch.

Frequently Asked Questions (FAQ)

Q: How is Sora Feed different from TikTok?

A: TikTok’s goal is to optimize watch time to retain users. Sora does the opposite, showing content most likely to inspire users to create their own videos first. It’s designed to focus on creation rather than consumption.

Q: What does it mean to adjust the algorithm with natural language?

A: Existing apps only recommend based on behavioral data like likes and watch time. Sora allows users to input text instructions like “Show me only sci-fi videos today” and the algorithm adjusts accordingly.

Q: Are there parental protection features?

A: Yes. Using ChatGPT parental control features, you can turn off feed personalization or limit continuous scrolling. Teen accounts have a default daily limit on videos they can create, and the Cameo feature (videos featuring other people) also has stricter permissions.


If you found this article useful, subscribe to AI Digester.

Reference Resources

Text→Image AI Training: FID Reduced by 30% Through This Method

Core Line 3: 200K step secret, Muon optimizer, token routing

  • REPA sort is only an early accelerator and should be removed after 200K steps
  • Muon optimizer alone achieves FID 18.2 → 15.55 (15% improvement)
  • At 1024×1024 high resolution, TREAD token routing reduces FID to 14.10

What happened?

The Photoroom team released an optimization guide for their text-to-image generation model PRX Part 2. [Hugging Face] While Part 1 covered architecture, this time they shared concrete ablation results on what to do during actual training.

Honestly, most technical documents of this kind end with “our model is the best,” but this is different. They also disclosed failed experiments and showed trade-offs of each technique with numbers.

Why is it important?

The cost of training a text-image model from scratch is enormous. A single wrong setting can waste thousands of GPU hours. The data released by Photoroom reduces such trial and error.

Personally, the most notable finding is about REPA (Representation Alignment). Using REPA-DINOv3 drops FID from 18.2 to 14.64. But there is a problem. Throughput decreases by 13% and training actually degrades after 200K steps. Simply put, it is only an early booster.

Another bug in BF16 weight storage. If you unknowingly save in BF16 instead of FP32, FID spikes from 18.2 to 21.87. That is an increase of 3.67. Surprisingly, many teams fall into this trap.

Practical Guide: Strategies by Resolution

Technique 256×256 FID 1024×1024 FID Throughput
Baseline 18.20 3.95 b/s
REPA-E-VAE 12.08 3.39 b/s
TREAD 21.61 ↑ 14.10 ↓ 1.64 b/s
Muon Optimizer 15.55

At 256×256, TREAD actually degrades quality. But at 1024×1024, completely different results are obtained. The higher the resolution, the greater the token routing effect.

What will happen in the future?

Photoroom will provide the complete training code in Part 3. They plan to release it and conduct a 24-hour “speed run.” The goal is to show how fast a good model can be built.

Personally, I think this release will have a significant impact on the open source image generation model ecosystem. Since Stable Diffusion, this is the first time training know-how has been disclosed in such detail.

Frequently Asked Questions (FAQ)

Q: When should REPA be removed?

A: After about 200K steps. It accelerates learning at first, but actually hinders convergence after that. This is clearly shown in Photoroom experiments. Missing the timing will degrade the quality of the final model.

Q: Should I use synthetic data or real images?

A: Use both. Use synthetic images at first to learn global structure, then use real images in later stages to capture high-frequency details. Using only compositing gives good FID but does not look photorealistic.

Q: How much better is Muon optimizer than AdamW?

A: About 15% improvement in FID. It drops from 18.2 to 15.55. Since computational cost is similar, there is no reason not to use it. However, hyperparameter tuning is slightly tricky.


If you found this article useful, please subscribe to AI Digester.

References

pi-mono: Claude Code Alternative AI Coding Agent 5.9k stars

pi-mono: Create Your Own AI Coding Agent in Your Terminal

  • GitHub Stars: 5.9k
  • Language: TypeScript 96.5%
  • License: MIT

Why This Project Is Popping Up

A developer felt Claude Code had become too complex. Mario Zechner experimented with LLM coding tools for 3 years and eventually decided to build his own tool.[Mario Zechner]

pi-mono is an AI agent toolkit built with the philosophy “don’t build it if you don’t need it.” It starts with a 1000-token system prompt and 4 core tools (read, write, edit, bash). Very lightweight compared to Claude Code’s thousands of token prompts. What’s in it?

  • Integrated LLM API: Use 15+ providers including OpenAI, Anthropic, Google, Azure, Mistral, Groq in one interface
  • Coding Agent CLI: Write, test, and debug code interactively in the terminal
  • Session Management: Pause and resume work, branch like git
  • Slack bot: Delegate Slack messages to the coding agent
  • vLLM pod management: Deploy and manage your own models on GPU pods
  • TUI/Web UI library: Build your own AI chat interface

Quick Start

# Install
npm install @mariozechner/pi-coding-agent

# run
npx pi

# or build from source
git clone https://github.com/badlogic/pi-mono
cd pi-mono
npm install && npm run build
./pi-test.sh

Where Can I Use It?

If Claude Code’s $200/month is too expensive and you prefer working in the terminal, pi could be an alternative. You only pay for API costs.

If you want to use self-hosted LLMs but existing tools don’t support them well, pi is the answer. It even has built-in vLLM pod management.

Personally, I think “transparency” is the biggest advantage. Claude Code runs invisible sub-agents internally to perform tasks. pi lets you directly see all model interactions.

Things to Note

  • Minimalism is the philosophy. MCP (Model Context Protocol) support is intentionally omitted
  • Full access called “YOLO mode” is the default. Be careful as permission checks are looser than Claude Code
  • Documentation is still lacking. Read the AGENTS.md file carefully

Similar Projects

Aider: Also an open source terminal coding tool. Similar in being model-agnostic, but pi covers a broader scope (UI library, pod management, etc.). [AIMultiple]

Claude Code: Has more features but requires a monthly subscription and has limitations on customization. pi allows freely adding features through TypeScript extensions.[Northflank]

Cursor: AI integrated into an IDE. If you prefer GUI over terminal, Cursor is better.

Frequently Asked Questions (FAQ)

Q: Can I use it for free?

A: pi is completely free under the MIT license. However, if you use external LLM APIs like OpenAI or Anthropic, those costs apply. You can use it without API costs by running Ollama or self-hosted vLLM locally.

Q: Is the performance good enough to replace Claude Code?

A: In Terminal-Bench 2.0 benchmarks, pi with Claude Opus 4.5 showed competitive results with Codex, Cursor, and Windsurf. This proves the minimalist approach doesn’t compromise performance.

Q: Does it support languages other than English?

A: The UI is in English, but if the connected LLM supports other languages, you can communicate and code in that language. You can write code with prompts in any language by connecting Claude or GPT-4.


If you found this article useful, please subscribe to AI Digester.

References

AI-Only SNS Moltbook: 17,000 Humans Behind 1.5 Million Bots

1.5 Million AI Agents, 17,000 Humans: The Hidden Truth

  • Moltbook, an AI-only SNS, has 1.5 million agents active, but only 17,000 humans behind them.
  • Wiz security team discovered a database vulnerability. 1.5 million API keys were exposed.
  • The founder admitted he “didn’t write a single line of code himself.” It was entirely ‘vibe-coded’ by an AI platform.

What Happened?

A security disaster struck Moltbook, a social network exclusively for AI agents. According to Wiz security team’s findings, behind 1.5 million AI agent accounts were only 17,000 humans. On average, each person was running 88 bots.[Wiz]

There’s an even more serious problem. Moltbook’s Supabase database was completely exposed. The API key was leaked in client-side JavaScript, and there were no Row Level Security policies at all. Anyone had read/write access to the entire database.[Axios]

The leaked information is shocking. It included 1.5 million API authentication tokens, 35,000 email addresses, and 4,060 private DMs between agents. In some conversations, OpenAI API keys were shared as plain text.[Techzine]

Why Is This Important?

Moltbook’s true nature has been revealed. The concept of an “autonomous social network of AI only” was actually closer to a puppet show controlled by humans behind the scenes.

Honestly, this was a disaster waiting to happen. As founder Matt Schlicht himself admitted, this platform was a ‘vibe-coded’ project, with the entire development left to an AI assistant “without writing a single line of code.” href=”https://www.engadget.com/ai/moltbook-the-ai-social-network-exposed-human-credentials-due-to-vibe-coded-security-flaw-230324567.html”>[Engadget] Security was naturally an afterthought.

Personally, I think this is a warning light for the AI agent era. Moltbook vividly showed how vulnerable security can be in systems where agents communicate with each other, process external data, and act autonomously.

Harlan Stewart of the Machine Intelligence Research Institute (MIRI) analyzed the viral screenshots and found that two-thirds were linked to human accounts marketing AI messaging apps.[Live Science]

What Happens Next?

Thanks to Wiz’s immediate report, the Moltbook team fixed the vulnerability within hours. But the fundamental problem remains unsolved.

AI agent expert Gary Marcus called Moltbook “a disaster waiting to happen.” AI models are simply recreating sci-fi scenarios from their training data. [Gary Marcus]

On the other hand, Andrej Karpathy called Moltbook “the most amazing sci-fi I’ve seen recently,” and Elon Musk called it “a very early stage of the singularity.” [Fortune]

But looking at it coolly, the current Moltbook is not evidence of AI autonomy, but evidence of how easily humans can manipulate AI systems.

Frequently Asked Questions

Q: What exactly is Moltbook?

A: An AI agents-only social network created by Matt Schlicht in January 2026. Similar in structure to Reddit, humans can only observe, and only AI agents like OpenClaw can write posts and comments. Currently over 1.5 million agents are registered.

Q: What is OpenClaw?

A: An open-source AI personal assistant software that runs locally on user devices. Originally launched as Clawdbot in November 2025, it was renamed to Moltbot due to a trademark request from Anthropic, then renamed again to OpenClaw in early 2026.

Q: Could my data have been leaked?

A: If you registered an OpenClaw agent on Moltbook, it’s possible. API keys, emails, and conversations between agents were exposed. Security researchers do not recommend using OpenClaw itself. Avoid it if you’re concerned about device security or data privacy.


If you found this article useful, subscribe to AI Digester.

References