From morning-ai
Defines unified specification for tracking, scoring, and validating AI news items in Product, Model, Benchmark, and Funding categories with include/exclude rules and source priorities.
npx claudepluginhub octo-patch/morningai --plugin morning-aiThis skill uses the workspace's default tool permissions.
This document defines the shared specification for scoring, formatting, and validating AI news items.
Searches for today's AI news across multiple categories, checks official changelogs, and creates/updates a briefing in Notion with deduplication against previously covered stories.
Generates concise Markdown text digest and 9:16 PNG image from scored AI news data for sharing on messaging platforms like WeChat, Telegram, Slack.
Monitors user-defined topics via scheduled web searches, AI importance scoring, proactive alerts, weekly digests, and memory-aware summaries. Use for tracking product releases, news, or tech updates.
Share bugs, ideas, or general feedback.
This document defines the shared specification for scoring, formatting, and validating AI news items.
This tracking system categorizes AI industry updates into 4 types. Each record must be labeled with its type. The main Agent can exclude unwanted types via exclude_types (All tracked by default).
Feature/version updates for AI tools and platforms.
| Include | Exclude |
|---|---|
| New feature launch (officially released) | Tips sharing (key people sharing usage tips, not new features) |
| Official version release (e.g. v2.0.0) | Pure marketing content (promotions, retweet giveaways) |
| Major capability upgrade (new model integration, new workflow) | Minor UI tweaks (interface changes not affecting functionality) |
| API/SDK update (new endpoints, new parameters) | Minor mobile update (bug fixes only) |
| Open-source project release (GitHub Releases) | |
| Product pricing change |
Source Priority: Changelog/Release Notes > GitHub Releases > Official Blog > Official X > Key People X
Key People Post Handling:
AI model releases, updates, and open-sourcing (including LLM, vision models, multimodal models).
| Include | Exclude |
|---|---|
| New model release (flagship model official launch) | App feature update (ChatGPT/Gemini interface features) |
| Model version update (series iteration versions) | Enterprise/Team product update |
| Model capability upgrade (context, multimodal) | Subscription change |
| Model API update (new endpoints, pricing adjustments) | Marketing campaigns, user milestones |
| Open-source model weights release (GitHub, HuggingFace) | |
| Multimodal model release (image, video, voice) | |
| Image/video generation model release |
Source Priority: Official Blog > Official X > API Changelog > GitHub/HuggingFace > Key People X > arXiv
Benchmark leaderboard changes, benchmark results, academic papers, technical reports.
| Include | Exclude |
|---|---|
| Official benchmark results | Survey papers (lacking novelty) |
| Authoritative leaderboard ranking changes (LMSYS, Artificial Analysis, VLM Arena) | Non-reproducible research |
| High-value academic papers (arXiv, top conferences) | Duplicate coverage (secondhand news restatements) |
| Major vendor technical reports/Research Blog | |
| Open-source research (papers with code/weights) | |
| Architecture innovation, training method breakthroughs | |
| Interpretability/alignment/safety research |
Source Priority: Benchmark institutions > arXiv > Official Research > HuggingFace Papers > KOL > Papers with Code > Reddit
Major funding, acquisitions/mergers, strategic partnerships, milestone events.
| Include | Exclude |
|---|---|
| Large funding (Series B+ or amount >= $100M) | Small seed/angel round (< $50M and Series A or below) |
| Acquisition/merger (AI-related companies) | Pure rumors/unconfirmed ("reportedly", "sources say") |
| Major strategic partnership (official integration with mainstream platforms) | Not directly AI-related acquisitions/funding |
| Strategic investment (involving AI companies) | Regular milestone (users < 1M) |
| Major milestone (users >= 1M / ARR >= $100M / DAU >= 1M) | Unilateral announcement, no response from counterparty |
Source Priority: Bilateral official confirmation > Crunchbase/PitchBook > Authoritative media (TechCrunch/The Verge)
Verification Rules:
| Event Type | Inclusion Criteria | Verification Method |
|---|---|---|
| Acquisition/merger | AI-related companies | Bilateral official confirmation or authoritative media coverage |
| Large funding | Series B+ or >= $100M | Official announcement or Crunchbase/PitchBook confirmation |
| Major milestone | Users >= 1M or ARR >= $100M | Official announcement or third-party data platform verification |
| Major partnership | Official integration with mainstream platforms | Both parties confirmed or feature subsequently launched |
| Priority | Source Type | Credibility | Handling |
|---|---|---|---|
| 1 | Official Blog/News | Highest | Accept directly |
| 2 | Official Changelog/Release Notes | Highest | Accept directly |
| 3 | Official X/Twitter | High | Accept directly |
| 4 | GitHub Releases | High | Accept directly |
| 5 | Key People X | Fairly High | Requires cross-verification |
| 6 | HuggingFace | Medium-High | Accept directly |
| 7 | arXiv | Medium-High | Accept directly |
| 8 | Benchmark institutions | Medium | Accept directly (for benchmark information) |
| 9 | Opinion Leaders/KOL | Reference | Must trace to official channel for confirmation |
| 10 | Industry Media | Reference | For lead discovery only, must trace to source |
Core Rule: Every specific number or technical detail in the report MUST be traceable to an authoritative primary source. Never infer, extrapolate, or "fill in" details from memory or general knowledge.
Specific details that MUST be verified from primary sources before inclusion:
| Detail Type | Authoritative Source | Example Error |
|---|---|---|
| Model parameter count | HuggingFace model card, official blog/paper | Writing "456B" when the actual size is "480B" |
| Model architecture (MoE active params, layers) | HuggingFace model card, technical report | Guessing active parameter count |
| Benchmark scores | Original benchmark site, official eval results | Citing an approximate score from memory |
| Version numbers | Official changelog, release notes, GitHub Release | Writing "v2.1" when the actual release is "v2.0" |
| Pricing | Official pricing page, API docs | Using outdated or incorrect price points |
| Context window size | Official documentation, model card | Confusing context lengths between model versions |
| Release/availability dates | Official announcement | Guessing a date based on general timeline |
| Funding amounts & valuations | Official press release, Crunchbase | Rounding or estimating funding figures |
| User counts / milestones | Official announcement, company blog | Using outdated user statistics |
| Training data details | Technical report, model card | Speculating about training data composition |
Verification protocol:
| Score | Verification Requirement |
|---|---|
| 7+ | Must have 2+ independent sources confirmed |
| 5-6 | Recommended 1+ other source corroboration |
| Below 5 | Single credible source sufficient |
For the following scenarios, a double-check is required (confirm the original event date through additional searches or official timelines):
"{product name} release date" or "{product name} announced" to confirmEvents where date cannot be confirmed → downgrade to skip, annotate "date cannot be confirmed"
Key Rule: must confirm the "actual event date", not "page accessible date" or "page last updated date".
A page being currently accessible does not mean the event it describes occurred within the window. Common misjudgment scenarios:
| Common Misjudgment Scenarios | Appearance | Correct Handling |
|---|---|---|
| Changelog page currently accessible | Page shows multiple historical entries | Check each entry's own date annotation |
| Leaderboard currently shows a model's ranking | Model currently on the list | Confirm the date the ranking change first occurred |
| Third-party platform page introduces a product | Page exists | Confirm the product's original release date, not the platform's listing date |
| News media reprints old news | Article publish date within window | Trace to original event date |
| Product website shows a feature | Feature currently available | Confirm the date the feature first launched/released |
Determination Process:
[time_window_start, time_window_end)[yesterday 08:00, today 08:00) UTC+8| Source | Timezone | Conversion |
|---|---|---|
| X/Twitter API | UTC | +8 hours → UTC+8 |
| Official Blog (US) | PST/PDT | +16/+15 hours → UTC+8 |
| Official Blog (China) | UTC+8 | No conversion needed |
| GitHub Releases | UTC | +8 hours → UTC+8 |
| Other sources | Case-by-case | Convert based on page annotation |
| Source Type | Correct Date Field | Incorrect Date Field | Notes |
|---|---|---|---|
| Changelog / Release Notes | Entry's own date annotation (e.g. "April 7, 2026") | Page access date, page last-modified | A Changelog page contains multiple historical records; must check entry date |
| Official Blog | Article header publish date (usually in URL or byline) | "last updated" or page footer copyright year | Distinguish between "publish date" and "last edited date" |
| GitHub Releases | Release's Published date | Repository's pushed_at or commit date | A repository having daily commits does not mean a new Release |
| X/Twitter | Tweet's created_at timestamp | — | Use directly, but requires UTC → UTC+8 conversion |
| Benchmark leaderboard | Date of first recorded ranking change | Current leaderboard access date | Model "currently on list" ≠ "just entered list"; check changelog to confirm change date |
| arXiv papers | Submitted / Announced date | Page access date | Note v1 submission date vs subsequent version update dates |
| News media | Original event date cited in the report | Article publish date | Media may report days after the event occurred |
| Product website | Feature/model's first release announcement date | Current page existence date | Product pages exist permanently; does not mean newly released |
| Third-party integration platform (fal.ai, Freepik, etc.) | Original release date of the integrated product | Platform listing/posting date | Platform "day 0 integration" means platform went live that day, but the product itself may have been released earlier |
Timeliness Check:
- ✅ Within window: Published Time ∈ [time_window_start, time_window_end)
- ❌ Outside window: Published Time < time_window_start or >= time_window_end
Morning-AI uses a two-stage scoring pipeline:
Stage 1 — Automated scoring (collect.py → lib/score.py):
Computes a 1-10 initial score from quantifiable metadata using 4 dimensions:
Stage 2 — Agent evaluation (report generation): The agent reviews each item using the 5 qualitative dimensions below. The agent may adjust the Stage 1 score based on content understanding — e.g., a low-engagement but groundbreaking paper might be scored up, while a viral but trivial post might be scored down.
| Dimension | Weight | Description |
|---|---|---|
| Impact | 30% | Industry impact of the event |
| Differentiation | 25% | Whether industry-first/unique |
| Breakthrough | 20% | Degree of technical/strategic breakthrough |
| Coverage | 15% | Affected users/scope |
| Timeliness | 10% | Time-sensitivity value of the information |
| Score | Level | Criteria |
|---|---|---|
| 9-10 | Major Event | Industry landscape breakthrough. Flagship model release, revolutionary feature, game-changing acquisition/partnership, unicorn-level funding ($1B+), record-breaking milestone |
| 7-8 | Important Update | Noteworthy important progress. Model series new version, major feature upgrade, official partnership with mainstream platform, large funding ($100M-$1B), major milestone (1M users/$100M ARR) |
| 5-6 | Regular Update | Routine updates worth noting. Minor version update, routine features, medium funding ($50M-$100M), general academic improvement |
| 3-4 | Minor Update | API parameter adjustments, doc updates, bug fixes, UI adjustments |
| 1-2 | Trivial Update | Typo fixes, dependency upgrades, detail optimization |
| Score | Criteria |
|---|---|
| 9-10 | Major vendor next-gen flagship model, industry landscape breakthrough |
| 7-8 | Model series new version, major capability improvement, important open-source model |
| 5-6 | Minor version update, API pricing adjustment, context extension |
| 3-4 | API parameter adjustments, doc updates |
| 1-2 | Bug fixes, detail optimization |
| Score | Criteria |
|---|---|
| 9-10 | Brand new major version, revolutionary feature, industry first |
| 7-8 | New model integration, major feature upgrade, core capability improvement |
| 5-6 | Routine feature addition, experience optimization |
| 3-4 | Bug fixes, UI adjustments, minor updates |
| 1-2 | Typo fixes, dependency upgrades |
| Score | Criteria |
|---|---|
| 9-10 | Paradigm-level breakthrough, potentially changing architecture design paradigm |
| 7-8 | High-value research, major technical innovation, with open-source code/weights |
| 5-6 | Valuable research, incremental improvement, validation experiments |
| 3-4 | Minor improvement, specific scenario optimization |
| 1-2 | Survey-type, lacking novelty, non-reproducible |
| Score | Criteria |
|---|---|
| 9-10 | Major company acquires well-known AI company, unicorn-level funding ($1B+), competition-changing partnership |
| 7-8 | Large funding ($100M-$1B), strategic acquisition, major milestone (1M users/$100M ARR) |
| 5-6 | Medium funding ($50M-$100M), general partnership |
| Factor | Positive Factors | Negative Factors |
|---|---|---|
| Impact scope | Industry-wide attention, official ecosystem support | Specific scenarios only |
| Technical breakthrough | First-of-kind, breakthrough, architecture innovation | Follow-up, catching up, routine iteration |
| Availability | Immediately available | Preview, waitlist |
| Open-source level | Weights open-sourced, code open-sourced | API only, closed use only |
| Strategic value | Major acquisition/funding, competition-changing | Internal optimization only |
### {Entity name} - {Event description}
| Field | Value |
|-------|-------|
| **Type Label** | Product / Model / Benchmark / Funding |
| **Timeliness Check** | ✅ Within window / ❌ Outside window |
| **Published Time** | YYYY-MM-DD HH:MM UTC+8 |
| **Event Type** | New feature / New model / Version update / Capability upgrade / Open-source release / Leaderboard change / Academic paper / Funding / Acquisition / Major partnership / Milestone / ... |
| **Partner/Acquirer** | (for Funding type) XX Company |
| **Amount/Scale** | (for Funding type) $XM / Series X / XM users |
| **Source** | [Source Name](URL) |
| **Score** | X.X |
**Summary**:
- Key point 1 (include specific details: version numbers, parameter counts, percentage improvements, pricing, availability)
- Key point 2 (competitive comparison or positioning)
- Key point 3 (technical specs or architecture details)
- Key point 4 (availability, rollout timeline, or ecosystem impact)
- Key point 5 (additional context as needed)
- (9-10 scores: 5-8 bullet points; 7-8 scores: 4-6 bullet points; 5-6 scores: 3-4 bullet points — cover all important aspects)
**Why It Matters** (required for 7+ scores):
> 1-4 sentence analysis of industry impact, competitive significance, or user implications. For 9-10 scores use 2-4 sentences with strategic context; for 7-8 scores use 1-2 sentences. Explain what this changes for the industry or end users — don't just restate what happened.
**Key Data** (required for 7+ scores when quantitative data exists — include when quantitative metrics are available):
| Metric | Value |
|--------|-------|
| e.g. Benchmark score | e.g. 92.3% (+5.1% vs previous SOTA) |
| e.g. Parameters | e.g. 671B total / 37B active |
| e.g. Pricing | e.g. $3/M input, $15/M output (vs $5/M previous) |
| e.g. Context length | e.g. 1M tokens (+4x vs v3) |
| e.g. Funding amount | e.g. $500M Series C at $5B valuation |
**Multi-source Verification** (required for 7+):
- [Source 1](URL)
- [Source 2](URL)
### {Entity name} - {Paper title}
| Field | Value |
|-------|-------|
| **Type Label** | Benchmark |
| **Timeliness Check** | ✅ Within window / ❌ Outside window |
| **Published Time** | YYYY-MM-DD HH:MM UTC+8 |
| **Event Type** | Academic paper / Technical report / Interpretability research |
| **arXiv** | [XXXX.XXXXX](https://arxiv.org/abs/XXXX.XXXXX) |
| **GitHub** | [Org/Repo](URL) (if available) |
| **Source** | [Source Name](URL) |
| **Score** | X.X |
**Core Innovation**:
> One-sentence description of the paper's core innovation
**Research Significance**:
> Who is impacted? What has changed? What does this enable that wasn't possible before? Include concrete implications for practitioners or downstream applications.
**Key Data**:
| Metric | Value |
|--------|-------|
| | |
| Entity | Summary | Skip Reason | Source |
|--------|---------|-------------|--------|
| [Entity name] | [Content description] | Pure marketing / Unconfirmed rumor / Outside window / Duplicate coverage / Tips sharing | [link](url) |
- **Entity** (X.X): Event description with specifics (version, capability, metric).
- Detail 1: what changed, key numbers, comparison with previous version or competitors
- Detail 2: additional context, availability, or technical specifics
- Detail 3: implications or notable aspects
Source: [Name](URL)
Use compact table format. The Source column must contain clickable [Name](URL) links.
| Entity | Score | Event | Source |
|--------|-------|-------|--------|
| Entity name | X.X | Brief description with one key detail | [Name](URL) |
FOR each source:
1️⃣ Check source (X account, Changelog/Blog, GitHub Releases, arXiv)
2️⃣ Timeliness check (per time validation rules)
3️⃣ Cross-verification (key people posts require cross-verification with official channels)
4️⃣ Content classification → determine type label (Product/Model/Benchmark/Funding)
5️⃣ Valid content → record with full format
6️⃣ Irrelevant content → record in skipped items with reason
END FOR
When a piece of information may belong to multiple types, classify by the following priority:
| Scenario | Classification |
|---|---|
| Product integrated new model | Product (core event is product feature change) |
| New model release brings product feature upgrade | Model (core event is model release) |
| Model leaderboard ranking change | Benchmark (core event is benchmark result) |
| Company received funding for model R&D | Funding (core event is funding) |
| Paper proposes new model architecture | Benchmark (academic papers fall under Benchmark) |