MAROKO133 Eksklusif ai: Anthropic's Claude Code can now read your Slack messages and

📌 MAROKO133 Breaking ai: Anthropic's Claude Code can now read your Slack mess

Anthropic on Monday launched a beta integration that connects its fast-growing Claude Code programming agent directly to Slack, allowing software engineers to delegate coding tasks without leaving the workplace messaging platform where much of their daily communication already happens.

The release, which Anthropic describes as a "research preview," is the AI safety company's latest move to embed its technology deeper into enterprise workflows — and comes as Claude Code has emerged as a surprise revenue engine, generating over $1 billion in annualized revenue just six months after its public debut in May.

"The critical context around engineering work often lives in Slack, including bug reports, feature requests, and engineering discussion," the company wrote in its announcement blog post. "When a bug report appears or a teammate needs a code fix, you can now tag Claude in Slack to automatically spin up a Claude Code session using the surrounding context."

From bug report to pull request: how the new Slack integration actually works

The mechanics are deceptively simple but address a persistent friction point in software development: the gap between where problems get discussed and where they get fixed.

When a user mentions @Claude in a Slack channel or thread, Claude analyzes the message to determine whether it constitutes a coding task. If it does, the system automatically creates a new Claude Code session. Users can also explicitly instruct Claude to treat requests as coding tasks.

Claude gathers context from recent channel and thread messages in Slack to feed into the Claude Code session. It will use this context to automatically choose which repository to run the task on based on the repositories you've authenticated to Claude Code on the web.

As the Claude Code session progresses, Claude posts status updates back to the Slack thread. Once complete, users receive a link to the full session where they can review changes, along with a direct link to open a pull request.

The feature builds on Anthropic's existing Claude for Slack integration and requires users to have access to Claude Code on the web. In practical terms, a product manager reporting a bug in Slack could tag Claude, which would then analyze the conversation context, identify the relevant code repository, investigate the issue, propose a fix, and post a pull request—all while updating the original Slack thread with its progress.

Why Anthropic is betting big on enterprise workflow integrations

The Slack integration arrives at a pivotal moment for Anthropic. Claude Code has already hit $1 billion in revenue six months since its public debut in May, according to a LinkedIn post from Anthropic's chief product officer, Mike Krieger. The coding agent continues to barrel toward scale with customers like Netflix, Spotify, and Salesforce.

The velocity of that growth helps explain why Anthropic made its first-ever acquisition earlier this month. Anthropic declined to comment on financial details. The Information earlier reported on Anthropic's bid to acquire Bun.

Bun is a breakthrough JavaScript runtime that is dramatically faster than the leading competition. As an all-in-one toolkit — combining runtime, package manager, bundler, and test runner — it's become essential infrastructure for AI-led software engineering, helping developers build and test applications at unprecedented velocity.

Since becoming generally available in May 2025, Claude Code has grown from its origins as an internal engineering experiment into a critical tool for many of the world's category-leading enterprises, including Netflix, Spotify, KPMG, L'Oreal, and Salesforce — and Bun has been key in helping scale its infrastructure throughout that evolution.

The acquisition signals that Anthropic views Claude Code not as a peripheral feature but as a core business line worth substantial investment. The Slack integration extends that bet, positioning Claude Code as an ambient presence in the workspaces where engineering decisions actually get made.

According to an Anthropic spokesperson, companies including Rakuten, Novo Nordisk, Uber, Snowflake, and Ramp now use Claude Code for both professional and novice developers. Rakuten, the Japanese e-commerce giant, has reportedly reduced software development timelines from 24 days to just 5 days using the tool — a 79% reduction that illustrates the productivity claims Anthropic has been making.

Claude Code's rapid rise from internal experiment to billion-dollar product

The Slack launch is the latest in a rapid series of Claude Code expansions. In late November, Claude Code was added to Anthropic's desktop apps including the Mac version. Claude Code was previously limited to mobile apps and the web. It allows software engineers to code, research, and update work with multiple local and remote sessions running at the same time.

That release accompanied Anthropic's unveiling of Claude Opus 4.5, its newest and most capable model. Claude Opus 4.5 is available today on the company's apps, API, and on all three major cloud platforms. Pricing is $5/$25 per million tokens — making Opus-level capabilities accessible to even more users, teams, and enterprises.

The company has also invested heavily in the developer infrastructure that powers Claude Code. In late November, Anth…

Konten dipersingkat otomatis.

đź”— Sumber: venturebeat.com


📌 MAROKO133 Hot ai: Z.ai debuts open source GLM-4.6V, a native tool-calling vision

Chinese AI startup Zhipu AI aka Z.ai has released its GLM-4.6V series, a new generation of open-source vision-language models (VLMs) optimized for multimodal reasoning, frontend automation, and high-efficiency deployment.

The release includes two models in "large" and "small" sizes:

  1. GLM-4.6V (106B), a larger 106-billion parameter model aimed at cloud-scale inference

  2. GLM-4.6V-Flash (9B), a smaller model of only 9 billion parameters designed for low-latency, local applications

Recall that generally speaking, models with more parameters — or internal settings governing their behavior, i.e. weights and biases — are more powerful, performant, and capable of performing at a higher general level across more varied tasks.

However, smaller models can offer better efficiency for edge or real-time applications where latency and resource constraints are critical.

The defining innovation in this series is the introduction of native function calling in a vision-language model—enabling direct use of tools such as search, cropping, or chart recognition with visual inputs.

With a 128,000 token context length (equivalent to a 300-page novel's worth of text exchanged in a single input/output interaction with the user) and state-of-the-art (SoTA) results across more than 20 benchmarks, the GLM-4.6V series positions itself as a highly competitive alternative to both closed and open-source VLMs. It's available in the following formats:

Licensing and Enterprise Use

GLM‑4.6V and GLM‑4.6V‑Flash are distributed under the MIT license, a permissive open-source license that allows free commercial and non-commercial use, modification, redistribution, and local deployment without obligation to open-source derivative works.

This licensing model makes the series suitable for enterprise adoption, including scenarios that require full control over infrastructure, compliance with internal governance, or air-gapped environments.

Model weights and documentation are publicly hosted on Hugging Face, with supporting code and tooling available on GitHub.

The MIT license ensures maximum flexibility for integration into proprietary systems, including internal tools, production pipelines, and edge deployments.

Architecture and Technical Capabilities

The GLM-4.6V models follow a conventional encoder-decoder architecture with significant adaptations for multimodal input.

Both models incorporate a Vision Transformer (ViT) encoder—based on AIMv2-Huge—and an MLP projector to align visual features with a large language model (LLM) decoder.

Video inputs benefit from 3D convolutions and temporal compression, while spatial encoding is handled using 2D-RoPE and bicubic interpolation of absolute positional embeddings.

A key technical feature is the system’s support for arbitrary image resolutions and aspect ratios, including wide panoramic inputs up to 200:1.

In addition to static image and document parsing, GLM-4.6V can ingest temporal sequences of video frames with explicit timestamp tokens, enabling robust temporal reasoning.

On the decoding side, the model supports token generation aligned with function-calling protocols, allowing for structured reasoning across text, image, and tool outputs. This is supported by extended tokenizer vocabulary and output formatting templates to ensure consistent API or agent compatibility.

Native Multimodal Tool Use

GLM-4.6V introduces native multimodal function calling, allowing visual assets—such as screenshots, images, and documents—to be passed directly as parameters to tools. This eliminates the need for intermediate text-only conversions, which have historically introduced information loss and complexity.

The tool invocation mechanism works bi-directionally:

  • Input tools can be passed images or videos directly (e.g., document pages to crop or analyze).

  • Output tools such as chart renderers or web snapshot utilities return visual data, which GLM-4.6V integrates directly into the reasoning chain.

In practice, this means GLM-4.6V can complete tasks such as:

  • Generating structured reports from mixed-format documents

  • Performing visual audit of candidate images

  • Automatically cropping figures from papers during generation

  • Conducting visual web search and answering multimodal queries

High Performance Benchmarks Compared to Other Similar-Sized Models

GLM-4.6V was evaluated across more than 20 public benchmarks covering general VQA, chart understanding, OCR, STEM reasoning, frontend replication, and multimodal agents.

According to the benchmark chart released by Zhipu AI:

  • GLM-4.6V (106B) achieves SoTA or near-SoTA scores among open-source models of comparable size (106B) on MMBench, MathVista, MMLongBench, ChartQAPro, RefCOCO, TreeBench, and more.

  • GLM-4.6V-Flash (9B) outperforms other lightweight models (e.g., Qwen3-VL-8B, GLM-4.1V-9B) across almost all categories tested.

  • The 106B model’s 128K-token window allows it to outperform larger models like Step-3 (321B) and Qwen3-VL-235B on long-context document tasks, video summarization, and structured multimodal reasoning.

Example scores from the leaderboard include:

  • MathVista: 88.2 (GLM-4.6V) vs. 84.6 (GLM-4.5V) vs. 81.4 (Qwen3-VL-8B)

  • WebVoyager: 81.0 vs. 68.4 (Qwen3-VL-8B)

  • Ref-L4-test: 88.9 vs. 89.5 (GLM-4.5V), but with better grounding fidelity at 87.7 (Flash) vs. 86.8

Both models were evaluated using the vLLM inference backend and support SGLang for video-based tasks.

Frontend Automation and Long-Context Workflows

Zhipu AI emphasized GLM-4.6V’s ability to support frontend development workflows. The model can:

  • Replicate pixel-accurate HTML/CSS/JS from UI screenshots

  • Accept natural language editing commands to modify layouts

  • Identify and manipulate specific UI components visually

This capability is integrated into an end-to-end visual programming interface, where the model iterates on layout, design intent, and output code using its native understanding of screen captures.

In long-document scenarios, GLM-4.6V can process up to 128,000 tokens—enabling a single inference pass across:

  • 150 pages of text (input)

  • 200 slide decks

  • 1-hour videos

Zhipu AI reported successful use of the model in financial analysis across multi-document corpora and in summarizing full-length sports broadcasts with timestamped event detection.

Training and Reinforcement Learning

The model was trained using multi-stage pre-training followed by supervised fine-tuning (SFT) and reinforcement learning (RL). Key innovations include:

  • Curriculum Sampling (RLCS): Dynamically adjusts the difficulty of training samples based on model progress

  • Multi-domain reward systems: Task-specific verifiers for STEM, chart reason…

    Konten dipersingkat otomatis.

    đź”— Sumber: venturebeat.com


🤖 Catatan MAROKO133

Artikel ini adalah rangkuman otomatis dari beberapa sumber terpercaya. Kami pilih topik yang sedang tren agar kamu selalu update tanpa ketinggalan.

✅ Update berikutnya dalam 30 menit — tema random menanti!

Author: timuna