发布日期:2026-03-07
收录条目:20
1. Microsoft Releases Phi-4-Reasoning-Vision-15B: A Compact Multimodal Model for Math, Science, and GUI Understanding
- 来源:MarkTechPost
- 发布时间:2026-03-06 23:53 UTC
- 链接:https://www.marktechpost.com/2026/03/06/microsoft-releases-phi-4-reasoning-vision-15b-a-compact-multimodal-model-for-math-science-and-gui-understanding/
摘要:Microsoft has released Phi-4-reasoning-vision-15B, a 15 billion parameter open-weight multimodal reasoning model designed for image and text tasks that require both perception and selective reasoning. It is a compact mod
2. A Production-Style NetworKit 11.2.1 Coding Tutorial for Large-Scale Graph Analytics, Communities, Cores, and Sparsification
- 来源:MarkTechPost
- 发布时间:2026-03-06 23:20 UTC
- 链接:https://www.marktechpost.com/2026/03/06/a-production-style-networkit-11-2-1-coding-tutorial-for-large-scale-graph-analytics-communities-cores-and-sparsification/
摘要:In this tutorial, we implement a production-grade, large-scale graph analytics pipeline in NetworKit, focusing on speed, memory efficiency, and version-safe APIs in NetworKit 11.2.1. We generate a large-scale free networ
3. Grammarly is using our identities without permission
- 来源:The Verge AI
- 发布时间:2026-03-06 20:58 UTC
- 链接:https://www.theverge.com/ai-artificial-intelligence/890921/grammarly-ai-expert-reviews
摘要:Grammarly's "expert review" feature offers to give users writing advice "inspired by" subject matter experts, including recently deceased professors, as Wired reported on Wednesday. When I tried the feature out myself, I
4. OpenAI Introduces Codex Security in Research Preview for Context-Aware Vulnerability Detection, Validation, and Patch Generation Across Codebases
- 来源:MarkTechPost
- 发布时间:2026-03-06 20:49 UTC
- 链接:https://www.marktechpost.com/2026/03/06/openai-introduces-codex-security-in-research-preview-for-context-aware-vulnerability-detection-validation-and-patch-generation-across-codebases/
摘要:OpenAI has introduced Codex Security, an application security agent that analyzes a codebase, validates likely vulnerabilities, and proposes fixes that developers can review before patching. The product is now rolling ou
5. Google AI Releases Android Bench: An Evaluation Framework and Leaderboard for LLMs in Android Development
- 来源:MarkTechPost
- 发布时间:2026-03-06 19:53 UTC
- 链接:https://www.marktechpost.com/2026/03/06/google-ai-releases-android-bench-an-evaluation-framework-and-leaderboard-for-llms-in-android-development/
摘要:Google has officially released Android Bench, a new leaderboard and evaluation framework designed to measure how Large Language Models (LLMs) perform specifically on Android development tasks. The dataset, methodology, a
6. The AI Doc is an overwrought hype piece for doomers and accelerationists alike
- 来源:The Verge AI
- 发布时间:2026-03-06 19:05 UTC
- 链接:https://www.theverge.com/entertainment/890806/the-ai-doc-or-how-i-became-an-apocaloptimist-review
摘要:We are in the thick of a massive push to incorporate generative AI into almost every aspect of our lives, but it is still easy to be confused about what it is and how it works. It doesn't help that many of gen AI's propo
7. Codex Security: now in research preview
- 来源:OpenAI News
- 发布时间:2026-03-06 10:00 UTC
- 链接:https://openai.com/index/codex-security-now-in-research-preview
摘要:Codex Security is an AI application security agent that analyzes project context to detect, validate, and patch complex vulnerabilities with higher confidence and less noise.
8. How Descript enables multilingual video dubbing at scale
- 来源:OpenAI News
- 发布时间:2026-03-06 10:00 UTC
- 链接:https://openai.com/index/descript
摘要:Descript uses OpenAI models to scale multilingual video dubbing, optimizing translations for both meaning and timing so dubbed speech sounds natural across languages.
9. How Balyasny Asset Management built an AI research engine for investing
- 来源:OpenAI News
- 发布时间:2026-03-06 07:00 UTC
- 链接:https://openai.com/index/balyasny-asset-management
摘要:See how Balyasny built an AI research system with GPT-5.4, rigorous model evaluation, and agent workflows to transform investment analysis at scale.
10. Liquid AI Releases LocalCowork Powered By LFM2-24B-A2B to Execute Privacy-First Agent Workflows Locally Via Model Context Protocol (MCP)
- 来源:MarkTechPost
- 发布时间:2026-03-06 05:45 UTC
- 链接:https://www.marktechpost.com/2026/03/05/liquid-ai-releases-localcowork-powered-by-lfm2-24b-a2b-to-execute-privacy-first-agent-workflows-locally-via-model-context-protocol-mcp/
摘要:Liquid AI has released LFM2-24B-A2B, a model optimized for local, low-latency tool dispatch, alongside LocalCowork, an open-source desktop agent application available in their Liquid4All GitHub Cookbook. The release prov
11. SkillNet: Create, Evaluate, and Connect AI Skills
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04448
摘要:arXiv:2603.04448v1 Announce Type: new Abstract: Current AI agents can flexibly invoke tools and execute complex tasks, yet their long-term advancement is hindered by the lack of systematic accumulation and transfer of sk
12. Capability Thresholds and Manufacturing Topology: How Embodied Intelligence Triggers Phase Transitions in Economic Geography
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04457
摘要:arXiv:2603.04457v1 Announce Type: new Abstract: The fundamental topology of manufacturing has not undergone a paradigm-level transformation since Henry Ford's moving assembly line in 1913. Every major innovation of the p
13. Progressive Refinement Regulation for Accelerating Diffusion Language Model Decoding
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04514
摘要:arXiv:2603.04514v1 Announce Type: new Abstract: Diffusion language models generate text through iterative denoising under a uniform refinement rule applied to all tokens. However, tokens stabilize at different rates in p
14. Discovering mathematical concepts through a multi-agent system
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04528
摘要:arXiv:2603.04528v1 Announce Type: new Abstract: Mathematical concepts emerge through an interplay of processes, including experimentation, efforts at proof, and counterexamples. In this paper, we present a new multi-agen
15. Adaptive Memory Admission Control for LLM Agents
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04549
摘要:arXiv:2603.04549v1 Announce Type: new Abstract: LLM-based agents increasingly rely on long-term memory to support multi-session reasoning and interaction, yet current systems provide little control over what information
16. Self-Attribution Bias: When AI Monitors Go Easy on Themselves
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04582
摘要:arXiv:2603.04582v1 Announce Type: new Abstract: Agentic systems increasingly rely on language models to monitor their own behavior. For example, coding agents may self critique generated code for pull request approval or
17. ECG-MoE: Mixture-of-Expert Electrocardiogram Foundation Model
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04589
摘要:arXiv:2603.04589v1 Announce Type: new Abstract: Electrocardiography (ECG) analysis is crucial for cardiac diagnosis, yet existing foundation models often fail to capture the periodicity and diverse features required for
18. Towards automated data analysis: A guided framework for LLM-based risk estimation
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04631
摘要:arXiv:2603.04631v1 Announce Type: new Abstract: Large Language Models (LLMs) are increasingly integrated into critical decision-making pipelines, a trend that raises the demand for robust and automated data analysis. Cur
19. When Agents Persuade: Propaganda Generation and Mitigation in LLMs
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04636
摘要:arXiv:2603.04636v1 Announce Type: new Abstract: Despite their wide-ranging benefits, LLM-based agents deployed in open environments can be exploited to produce manipulative material. In this study, we task LLMs with prop
20. Using Vision + Language Models to Predict Item Difficulty
- 来源:arXiv cs.AI
- 发布时间:2026-03-06 05:00 UTC
- 链接:https://arxiv.org/abs/2603.04670
摘要:arXiv:2603.04670v1 Announce Type: new Abstract: This project investigates the capabilities of large language models (LLMs) to determine the difficulty of data visualization literacy test items. We explore whether feature