@wentorai/research-plugins 1.2.2 → 1.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +16 -8
- package/openclaw.plugin.json +10 -3
- package/package.json +2 -5
- package/skills/analysis/dataviz/SKILL.md +25 -0
- package/skills/analysis/dataviz/chart-image-generator/SKILL.md +1 -1
- package/skills/analysis/econometrics/SKILL.md +23 -0
- package/skills/analysis/econometrics/robustness-checks/SKILL.md +1 -1
- package/skills/analysis/statistics/SKILL.md +21 -0
- package/skills/analysis/statistics/data-anomaly-detection/SKILL.md +1 -1
- package/skills/analysis/statistics/ml-experiment-tracker/SKILL.md +1 -1
- package/skills/analysis/statistics/{senior-data-scientist-guide → modeling-strategy-guide}/SKILL.md +5 -5
- package/skills/analysis/wrangling/SKILL.md +21 -0
- package/skills/analysis/wrangling/csv-data-analyzer/SKILL.md +1 -1
- package/skills/analysis/wrangling/data-cog-guide/SKILL.md +1 -1
- package/skills/domains/ai-ml/SKILL.md +37 -0
- package/skills/domains/biomedical/SKILL.md +28 -0
- package/skills/domains/biomedical/genomas-guide/SKILL.md +1 -1
- package/skills/domains/biomedical/med-researcher-guide/SKILL.md +1 -1
- package/skills/domains/biomedical/medgeclaw-guide/SKILL.md +1 -1
- package/skills/domains/business/SKILL.md +17 -0
- package/skills/domains/business/architecture-design-guide/SKILL.md +1 -1
- package/skills/domains/chemistry/SKILL.md +19 -0
- package/skills/domains/chemistry/computational-chemistry-guide/SKILL.md +1 -1
- package/skills/domains/cs/SKILL.md +21 -0
- package/skills/domains/ecology/SKILL.md +16 -0
- package/skills/domains/economics/SKILL.md +20 -0
- package/skills/domains/economics/post-labor-economics/SKILL.md +1 -1
- package/skills/domains/economics/pricing-psychology-guide/SKILL.md +1 -1
- package/skills/domains/education/SKILL.md +19 -0
- package/skills/domains/education/academic-study-methods/SKILL.md +1 -1
- package/skills/domains/education/edumcp-guide/SKILL.md +1 -1
- package/skills/domains/finance/SKILL.md +19 -0
- package/skills/domains/finance/akshare-finance-data/SKILL.md +1 -1
- package/skills/domains/finance/options-analytics-agent-guide/SKILL.md +1 -1
- package/skills/domains/finance/stata-accounting-research/SKILL.md +1 -1
- package/skills/domains/geoscience/SKILL.md +17 -0
- package/skills/domains/humanities/SKILL.md +16 -0
- package/skills/domains/humanities/history-research-guide/SKILL.md +1 -1
- package/skills/domains/humanities/political-history-guide/SKILL.md +1 -1
- package/skills/domains/law/SKILL.md +19 -0
- package/skills/domains/math/SKILL.md +17 -0
- package/skills/domains/pharma/SKILL.md +17 -0
- package/skills/domains/physics/SKILL.md +16 -0
- package/skills/domains/social-science/SKILL.md +17 -0
- package/skills/domains/social-science/sociology-research-methods/SKILL.md +1 -1
- package/skills/literature/discovery/SKILL.md +20 -0
- package/skills/literature/discovery/paper-recommendation-guide/SKILL.md +1 -1
- package/skills/literature/discovery/semantic-paper-radar/SKILL.md +1 -1
- package/skills/literature/fulltext/SKILL.md +26 -0
- package/skills/literature/metadata/SKILL.md +35 -0
- package/skills/literature/metadata/doi-content-negotiation/SKILL.md +4 -0
- package/skills/literature/metadata/doi-resolution-guide/SKILL.md +4 -0
- package/skills/literature/metadata/orcid-api/SKILL.md +4 -0
- package/skills/literature/metadata/orcid-integration-guide/SKILL.md +4 -0
- package/skills/literature/search/SKILL.md +43 -0
- package/skills/literature/search/paper-search-mcp-guide/SKILL.md +1 -1
- package/skills/research/automation/SKILL.md +21 -0
- package/skills/research/deep-research/SKILL.md +24 -0
- package/skills/research/deep-research/auto-deep-research-guide/SKILL.md +1 -1
- package/skills/research/deep-research/in-depth-research-guide/SKILL.md +1 -1
- package/skills/research/funding/SKILL.md +20 -0
- package/skills/research/methodology/SKILL.md +24 -0
- package/skills/research/paper-review/SKILL.md +19 -0
- package/skills/research/paper-review/paper-critique-framework/SKILL.md +1 -1
- package/skills/tools/code-exec/SKILL.md +18 -0
- package/skills/tools/diagram/SKILL.md +20 -0
- package/skills/tools/document/SKILL.md +21 -0
- package/skills/tools/knowledge-graph/SKILL.md +21 -0
- package/skills/tools/ocr-translate/SKILL.md +18 -0
- package/skills/tools/ocr-translate/handwriting-recognition-guide/SKILL.md +2 -0
- package/skills/tools/ocr-translate/latex-ocr-guide/SKILL.md +2 -0
- package/skills/tools/scraping/SKILL.md +17 -0
- package/skills/writing/citation/SKILL.md +33 -0
- package/skills/writing/citation/zotfile-attachment-guide/SKILL.md +2 -0
- package/skills/writing/composition/SKILL.md +22 -0
- package/skills/writing/composition/research-paper-writer/SKILL.md +1 -1
- package/skills/writing/composition/scientific-writing-wrapper/SKILL.md +1 -1
- package/skills/writing/latex/SKILL.md +22 -0
- package/skills/writing/latex/academic-writing-latex/SKILL.md +1 -1
- package/skills/writing/latex/latex-drawing-guide/SKILL.md +1 -1
- package/skills/writing/polish/SKILL.md +20 -0
- package/skills/writing/polish/chinese-text-humanizer/SKILL.md +1 -1
- package/skills/writing/templates/SKILL.md +22 -0
- package/skills/writing/templates/beamer-presentation-guide/SKILL.md +1 -1
- package/skills/writing/templates/scientific-article-pdf/SKILL.md +1 -1
- package/skills/analysis/dataviz/citation-map-guide/SKILL.md +0 -184
- package/skills/analysis/dataviz/data-visualization-principles/SKILL.md +0 -171
- package/skills/analysis/econometrics/empirical-paper-analysis/SKILL.md +0 -192
- package/skills/analysis/econometrics/panel-data-regression-workflow/SKILL.md +0 -267
- package/skills/analysis/econometrics/stata-regression/SKILL.md +0 -117
- package/skills/analysis/statistics/general-statistics-guide/SKILL.md +0 -226
- package/skills/analysis/statistics/infiagent-benchmark-guide/SKILL.md +0 -106
- package/skills/analysis/statistics/pywayne-statistics-guide/SKILL.md +0 -192
- package/skills/analysis/statistics/quantitative-methods-guide/SKILL.md +0 -193
- package/skills/analysis/wrangling/claude-data-analysis-guide/SKILL.md +0 -100
- package/skills/analysis/wrangling/open-data-scientist-guide/SKILL.md +0 -197
- package/skills/domains/ai-ml/annotated-dl-papers-guide/SKILL.md +0 -159
- package/skills/domains/humanities/digital-humanities-methods/SKILL.md +0 -232
- package/skills/domains/law/legal-research-methods/SKILL.md +0 -190
- package/skills/domains/social-science/sociology-research-guide/SKILL.md +0 -238
- package/skills/literature/discovery/arxiv-paper-monitoring/SKILL.md +0 -233
- package/skills/literature/discovery/paper-tracking-guide/SKILL.md +0 -211
- package/skills/literature/fulltext/zotero-scihub-guide/SKILL.md +0 -168
- package/skills/literature/search/arxiv-osiris/SKILL.md +0 -199
- package/skills/literature/search/deepgit-search-guide/SKILL.md +0 -147
- package/skills/literature/search/multi-database-literature-search/SKILL.md +0 -198
- package/skills/literature/search/papers-chat-guide/SKILL.md +0 -194
- package/skills/literature/search/pasa-paper-search-guide/SKILL.md +0 -138
- package/skills/literature/search/scientify-literature-survey/SKILL.md +0 -203
- package/skills/research/automation/ai-scientist-guide/SKILL.md +0 -228
- package/skills/research/automation/coexist-ai-guide/SKILL.md +0 -149
- package/skills/research/automation/foam-agent-guide/SKILL.md +0 -203
- package/skills/research/automation/research-paper-orchestrator/SKILL.md +0 -254
- package/skills/research/deep-research/academic-deep-research/SKILL.md +0 -190
- package/skills/research/deep-research/cognitive-kernel-guide/SKILL.md +0 -200
- package/skills/research/deep-research/corvus-research-guide/SKILL.md +0 -132
- package/skills/research/deep-research/deep-research-pro/SKILL.md +0 -213
- package/skills/research/deep-research/deep-research-work/SKILL.md +0 -204
- package/skills/research/deep-research/research-cog/SKILL.md +0 -153
- package/skills/research/methodology/academic-mentor-guide/SKILL.md +0 -169
- package/skills/research/methodology/deep-innovator-guide/SKILL.md +0 -242
- package/skills/research/methodology/research-pipeline-units-guide/SKILL.md +0 -169
- package/skills/research/paper-review/paper-compare-guide/SKILL.md +0 -238
- package/skills/research/paper-review/paper-digest-guide/SKILL.md +0 -240
- package/skills/research/paper-review/paper-research-assistant/SKILL.md +0 -231
- package/skills/research/paper-review/research-quality-filter/SKILL.md +0 -261
- package/skills/tools/code-exec/contextplus-mcp-guide/SKILL.md +0 -110
- package/skills/tools/diagram/clawphd-guide/SKILL.md +0 -149
- package/skills/tools/diagram/scientific-graphical-abstract/SKILL.md +0 -201
- package/skills/tools/document/md2pdf-xelatex/SKILL.md +0 -212
- package/skills/tools/document/openpaper-guide/SKILL.md +0 -232
- package/skills/tools/document/weknora-guide/SKILL.md +0 -216
- package/skills/tools/knowledge-graph/mimir-memory-guide/SKILL.md +0 -135
- package/skills/tools/knowledge-graph/open-webui-tools-guide/SKILL.md +0 -156
- package/skills/tools/ocr-translate/formula-recognition-guide/SKILL.md +0 -367
- package/skills/tools/ocr-translate/math-equation-renderer/SKILL.md +0 -198
- package/skills/tools/scraping/api-data-collection-guide/SKILL.md +0 -301
- package/skills/writing/citation/academic-citation-manager-guide/SKILL.md +0 -182
- package/skills/writing/composition/opendraft-thesis-guide/SKILL.md +0 -200
- package/skills/writing/composition/paper-debugger-guide/SKILL.md +0 -143
- package/skills/writing/composition/paperforge-guide/SKILL.md +0 -205
|
@@ -1,132 +0,0 @@
|
|
|
1
|
-
---
|
|
2
|
-
name: corvus-research-guide
|
|
3
|
-
description: "Multi-agent AI research with semantic search and citation snowballing"
|
|
4
|
-
metadata:
|
|
5
|
-
openclaw:
|
|
6
|
-
emoji: "🐦⬛"
|
|
7
|
-
category: "research"
|
|
8
|
-
subcategory: "deep-research"
|
|
9
|
-
keywords: ["Corvus", "multi-agent", "semantic search", "citation snowballing", "research synthesis", "AI research"]
|
|
10
|
-
source: "https://github.com/corvus-research/corvus"
|
|
11
|
-
---
|
|
12
|
-
|
|
13
|
-
# Corvus Research Guide
|
|
14
|
-
|
|
15
|
-
## Overview
|
|
16
|
-
|
|
17
|
-
Corvus is a multi-agent AI research system that combines semantic search, forward/backward citation snowballing, and synthesis to conduct thorough literature investigations. It iteratively expands search results by following citation chains, identifies research gaps, and generates structured research briefs with full provenance.
|
|
18
|
-
|
|
19
|
-
## Architecture
|
|
20
|
-
|
|
21
|
-
### Agent Pipeline
|
|
22
|
-
|
|
23
|
-
```
|
|
24
|
-
Query → Semantic Search Agent
|
|
25
|
-
↓
|
|
26
|
-
Citation Snowball Agent (forward + backward)
|
|
27
|
-
↓
|
|
28
|
-
Relevance Filter Agent
|
|
29
|
-
↓
|
|
30
|
-
Synthesis Agent
|
|
31
|
-
↓
|
|
32
|
-
Report with Citation Graph
|
|
33
|
-
```
|
|
34
|
-
|
|
35
|
-
### Key Features
|
|
36
|
-
|
|
37
|
-
1. **Semantic search**: Uses embedding-based search across Semantic Scholar, OpenAlex
|
|
38
|
-
2. **Citation snowballing**: Iteratively follows references (backward) and citations (forward) to discover related work
|
|
39
|
-
3. **Relevance scoring**: AI-based relevance assessment at each expansion step
|
|
40
|
-
4. **Provenance tracking**: Every claim linked to source papers
|
|
41
|
-
5. **Gap identification**: Identifies under-explored research areas
|
|
42
|
-
|
|
43
|
-
## Usage
|
|
44
|
-
|
|
45
|
-
```python
|
|
46
|
-
from corvus import ResearchAgent
|
|
47
|
-
|
|
48
|
-
agent = ResearchAgent(
|
|
49
|
-
llm_provider="anthropic",
|
|
50
|
-
search_backends=["semantic_scholar", "openalex"],
|
|
51
|
-
max_snowball_depth=2,
|
|
52
|
-
)
|
|
53
|
-
|
|
54
|
-
# Conduct deep research
|
|
55
|
-
result = agent.research(
|
|
56
|
-
query="What are the current approaches to continual learning "
|
|
57
|
-
"in large language models?",
|
|
58
|
-
initial_papers=20,
|
|
59
|
-
snowball_per_paper=5,
|
|
60
|
-
)
|
|
61
|
-
|
|
62
|
-
# Access results
|
|
63
|
-
print(f"Papers found: {len(result.papers)}")
|
|
64
|
-
print(f"Unique clusters: {len(result.clusters)}")
|
|
65
|
-
print(f"\nSynthesis:\n{result.synthesis}")
|
|
66
|
-
|
|
67
|
-
# Export citation graph
|
|
68
|
-
result.export_graph("citation_network.gexf")
|
|
69
|
-
|
|
70
|
-
# Export bibliography
|
|
71
|
-
result.export_bibtex("references.bib")
|
|
72
|
-
```
|
|
73
|
-
|
|
74
|
-
### Snowballing Configuration
|
|
75
|
-
|
|
76
|
-
```python
|
|
77
|
-
agent = ResearchAgent(
|
|
78
|
-
snowball_config={
|
|
79
|
-
"max_depth": 3, # Citation chain depth
|
|
80
|
-
"backward_limit": 10, # References per paper
|
|
81
|
-
"forward_limit": 10, # Citations per paper
|
|
82
|
-
"relevance_threshold": 0.7, # Min relevance to continue
|
|
83
|
-
"year_filter": 2020, # Only papers from 2020+
|
|
84
|
-
}
|
|
85
|
-
)
|
|
86
|
-
```
|
|
87
|
-
|
|
88
|
-
### Research Modes
|
|
89
|
-
|
|
90
|
-
```python
|
|
91
|
-
# Broad survey mode
|
|
92
|
-
result = agent.research(query, mode="survey",
|
|
93
|
-
initial_papers=50)
|
|
94
|
-
|
|
95
|
-
# Focused deep-dive
|
|
96
|
-
result = agent.research(query, mode="focused",
|
|
97
|
-
initial_papers=10,
|
|
98
|
-
snowball_depth=3)
|
|
99
|
-
|
|
100
|
-
# Gap analysis
|
|
101
|
-
result = agent.research(query, mode="gap_analysis")
|
|
102
|
-
# Returns underexplored subtopics and suggested directions
|
|
103
|
-
```
|
|
104
|
-
|
|
105
|
-
## Output Format
|
|
106
|
-
|
|
107
|
-
```python
|
|
108
|
-
# Structured research brief
|
|
109
|
-
brief = result.generate_brief()
|
|
110
|
-
|
|
111
|
-
# Contains:
|
|
112
|
-
# - Research question
|
|
113
|
-
# - Methodology (search strategy, databases, snowball depth)
|
|
114
|
-
# - Key themes (clustered by topic)
|
|
115
|
-
# - Timeline (research evolution over time)
|
|
116
|
-
# - Gap analysis (underexplored areas)
|
|
117
|
-
# - Bibliography (all papers with citation counts)
|
|
118
|
-
|
|
119
|
-
brief.save("research_brief.md")
|
|
120
|
-
```
|
|
121
|
-
|
|
122
|
-
## Use Cases
|
|
123
|
-
|
|
124
|
-
1. **Literature reviews**: Comprehensive coverage via snowballing
|
|
125
|
-
2. **Research gap identification**: Find underexplored subtopics
|
|
126
|
-
3. **Trend analysis**: Track research evolution through citation chains
|
|
127
|
-
4. **Grant proposals**: Quick evidence of research need
|
|
128
|
-
|
|
129
|
-
## References
|
|
130
|
-
|
|
131
|
-
- [Corvus GitHub](https://github.com/corvus-research/corvus)
|
|
132
|
-
- Wohlin, C. (2014). "Guidelines for snowballing in systematic literature studies." *EASE 2014*.
|
|
@@ -1,213 +0,0 @@
|
|
|
1
|
-
---
|
|
2
|
-
name: deep-research-pro
|
|
3
|
-
description: "Multi-source deep research agent with verification and synthesis"
|
|
4
|
-
metadata:
|
|
5
|
-
openclaw:
|
|
6
|
-
emoji: "🕵️"
|
|
7
|
-
category: "research"
|
|
8
|
-
subcategory: "deep-research"
|
|
9
|
-
keywords: ["deep research", "multi-source", "source verification", "research agent", "evidence synthesis", "fact checking"]
|
|
10
|
-
source: "https://github.com/AcademicSkills/deep-research-pro"
|
|
11
|
-
---
|
|
12
|
-
|
|
13
|
-
# Deep Research Pro
|
|
14
|
-
|
|
15
|
-
A professional-grade deep research methodology that coordinates multi-source information gathering, cross-reference verification, and structured synthesis. Designed for research tasks that require high confidence in factual accuracy, comprehensive coverage, and traceable evidence chains.
|
|
16
|
-
|
|
17
|
-
## Overview
|
|
18
|
-
|
|
19
|
-
Deep Research Pro implements an agent-based methodology where the research process is decomposed into specialized phases: query decomposition, parallel source gathering, cross-reference verification, contradiction resolution, and structured synthesis. Unlike simple search-and-summarize approaches, this skill emphasizes source triangulation, evidence grading, and explicit uncertainty marking.
|
|
20
|
-
|
|
21
|
-
The methodology is particularly valuable for literature reviews, technology assessments, policy analyses, and any research task where decision-makers need to trust the completeness and accuracy of the findings. Every claim in the final output is linked to at least one verified source, with confidence levels assigned based on the quality and agreement of the evidence.
|
|
22
|
-
|
|
23
|
-
## Research Agent Architecture
|
|
24
|
-
|
|
25
|
-
### Phase 1: Query Decomposition
|
|
26
|
-
|
|
27
|
-
```python
|
|
28
|
-
def decompose_research_query(query: str) -> dict:
|
|
29
|
-
"""
|
|
30
|
-
Break a complex research question into atomic sub-questions
|
|
31
|
-
that can be independently investigated.
|
|
32
|
-
|
|
33
|
-
Example:
|
|
34
|
-
Input: "What is the current state of quantum computing for
|
|
35
|
-
drug discovery, and which companies are leading?"
|
|
36
|
-
|
|
37
|
-
Output: {
|
|
38
|
-
'sub_questions': [
|
|
39
|
-
'What quantum computing methods are used in drug discovery?',
|
|
40
|
-
'What are the computational advantages over classical methods?',
|
|
41
|
-
'Which companies are actively working on quantum drug discovery?',
|
|
42
|
-
'What drugs or molecules have been studied using quantum methods?',
|
|
43
|
-
'What are the current limitations and timeline to practical use?'
|
|
44
|
-
],
|
|
45
|
-
'cross_cutting_themes': [
|
|
46
|
-
'Cost-benefit vs classical computing',
|
|
47
|
-
'Regulatory considerations',
|
|
48
|
-
'Academic vs industry progress'
|
|
49
|
-
],
|
|
50
|
-
'source_strategy': {
|
|
51
|
-
'academic': ['PubMed', 'arXiv', 'Google Scholar'],
|
|
52
|
-
'industry': ['company websites', 'press releases', 'patent databases'],
|
|
53
|
-
'grey': ['government reports', 'consulting firm analyses']
|
|
54
|
-
}
|
|
55
|
-
}
|
|
56
|
-
"""
|
|
57
|
-
pass # Implemented by the agent's reasoning
|
|
58
|
-
```
|
|
59
|
-
|
|
60
|
-
### Phase 2: Parallel Source Gathering
|
|
61
|
-
|
|
62
|
-
For each sub-question, gather evidence from multiple independent source types:
|
|
63
|
-
|
|
64
|
-
| Source Type | Examples | Strengths | Limitations |
|
|
65
|
-
|------------|---------|-----------|-------------|
|
|
66
|
-
| **Peer-reviewed** | Journal articles, conference papers | Rigorous review | Publication lag |
|
|
67
|
-
| **Preprints** | arXiv, bioRxiv, SSRN | Current | Not peer-reviewed |
|
|
68
|
-
| **Institutional** | WHO, NIH, government agencies | Authoritative | May be conservative |
|
|
69
|
-
| **Industry** | Company blogs, press releases | Current, practical | Biased toward own products |
|
|
70
|
-
| **News/Media** | Science journalism, trade publications | Accessible | May oversimplify |
|
|
71
|
-
| **Expert opinion** | Interviews, blog posts, talks | Nuanced | Subjective |
|
|
72
|
-
|
|
73
|
-
### Phase 3: Source Verification
|
|
74
|
-
|
|
75
|
-
```python
|
|
76
|
-
def verify_source(source: dict) -> dict:
|
|
77
|
-
"""
|
|
78
|
-
Assess the credibility and reliability of a single source.
|
|
79
|
-
"""
|
|
80
|
-
verification = {
|
|
81
|
-
'source_id': source['id'],
|
|
82
|
-
'url': source['url'],
|
|
83
|
-
'checks': {}
|
|
84
|
-
}
|
|
85
|
-
|
|
86
|
-
# Check 1: Author credibility
|
|
87
|
-
verification['checks']['author'] = {
|
|
88
|
-
'identifiable': bool(source.get('author')),
|
|
89
|
-
'affiliated': bool(source.get('institution')),
|
|
90
|
-
'h_index_available': bool(source.get('h_index')),
|
|
91
|
-
'domain_expert': source.get('expertise_match', False)
|
|
92
|
-
}
|
|
93
|
-
|
|
94
|
-
# Check 2: Publication venue
|
|
95
|
-
verification['checks']['venue'] = {
|
|
96
|
-
'peer_reviewed': source.get('peer_reviewed', False),
|
|
97
|
-
'impact_factor': source.get('impact_factor'),
|
|
98
|
-
'known_predatory': source.get('predatory_flag', False)
|
|
99
|
-
}
|
|
100
|
-
|
|
101
|
-
# Check 3: Currency
|
|
102
|
-
verification['checks']['currency'] = {
|
|
103
|
-
'publication_year': source.get('year'),
|
|
104
|
-
'is_recent': source.get('year', 0) >= 2023,
|
|
105
|
-
'superseded': source.get('retracted', False)
|
|
106
|
-
}
|
|
107
|
-
|
|
108
|
-
# Check 4: Corroboration
|
|
109
|
-
verification['checks']['corroboration'] = {
|
|
110
|
-
'cited_by_count': source.get('citation_count', 0),
|
|
111
|
-
'independent_confirmation': source.get('replicated', False),
|
|
112
|
-
'consistent_with_consensus': source.get('consensus_aligned', None)
|
|
113
|
-
}
|
|
114
|
-
|
|
115
|
-
# Overall confidence score
|
|
116
|
-
score = sum([
|
|
117
|
-
verification['checks']['author']['identifiable'],
|
|
118
|
-
verification['checks']['venue']['peer_reviewed'],
|
|
119
|
-
not verification['checks']['venue']['known_predatory'],
|
|
120
|
-
verification['checks']['currency']['is_recent'],
|
|
121
|
-
verification['checks']['corroboration']['cited_by_count'] > 5
|
|
122
|
-
]) / 5.0
|
|
123
|
-
|
|
124
|
-
verification['confidence'] = round(score, 2)
|
|
125
|
-
verification['grade'] = (
|
|
126
|
-
'A' if score >= 0.8 else
|
|
127
|
-
'B' if score >= 0.6 else
|
|
128
|
-
'C' if score >= 0.4 else 'D'
|
|
129
|
-
)
|
|
130
|
-
return verification
|
|
131
|
-
```
|
|
132
|
-
|
|
133
|
-
## Cross-Reference and Contradiction Resolution
|
|
134
|
-
|
|
135
|
-
### Triangulation Protocol
|
|
136
|
-
|
|
137
|
-
For each factual claim in the research output:
|
|
138
|
-
|
|
139
|
-
1. **Identify the claim**: Extract a specific, testable statement.
|
|
140
|
-
2. **Find supporting sources**: At least 2 independent sources.
|
|
141
|
-
3. **Check for contradicting sources**: Actively search for disconfirming evidence.
|
|
142
|
-
4. **Assess source quality**: Weight evidence by source grade (A > B > C > D).
|
|
143
|
-
5. **Assign confidence level**: Based on evidence agreement and quality.
|
|
144
|
-
|
|
145
|
-
```
|
|
146
|
-
Confidence Levels:
|
|
147
|
-
HIGH: 3+ Grade A/B sources agree, no contradictions
|
|
148
|
-
MEDIUM: 2+ sources agree, minor contradictions or lower-quality sources
|
|
149
|
-
LOW: Single source, or significant contradictions among sources
|
|
150
|
-
UNCERTAIN: Conflicting evidence of similar quality; state both positions
|
|
151
|
-
```
|
|
152
|
-
|
|
153
|
-
### Contradiction Handling
|
|
154
|
-
|
|
155
|
-
When sources disagree:
|
|
156
|
-
|
|
157
|
-
1. Check if the disagreement is due to different definitions, timeframes, or contexts.
|
|
158
|
-
2. Check if one source has been superseded or corrected.
|
|
159
|
-
3. If genuine disagreement exists, present both positions with their evidence.
|
|
160
|
-
4. Note the disagreement explicitly: "Source A (Grade B, 2024) reports X, while Source B (Grade A, 2023) reports Y. The discrepancy may be due to [possible explanation]."
|
|
161
|
-
|
|
162
|
-
## Structured Synthesis Output
|
|
163
|
-
|
|
164
|
-
### Report Template
|
|
165
|
-
|
|
166
|
-
```markdown
|
|
167
|
-
# Research Report: [Topic]
|
|
168
|
-
|
|
169
|
-
## Executive Summary
|
|
170
|
-
[3-5 sentence overview of key findings with confidence levels]
|
|
171
|
-
|
|
172
|
-
## Methodology
|
|
173
|
-
- Sub-questions investigated: [list]
|
|
174
|
-
- Sources consulted: [count by type]
|
|
175
|
-
- Date range of sources: [range]
|
|
176
|
-
- Verification standard: [triangulation with Grade B+ sources]
|
|
177
|
-
|
|
178
|
-
## Findings
|
|
179
|
-
|
|
180
|
-
### Finding 1: [Title]
|
|
181
|
-
**Confidence: HIGH**
|
|
182
|
-
[Description with inline source references]
|
|
183
|
-
Sources: [Source1 (Grade A)], [Source2 (Grade B)], [Source3 (Grade A)]
|
|
184
|
-
|
|
185
|
-
### Finding 2: [Title]
|
|
186
|
-
**Confidence: MEDIUM**
|
|
187
|
-
[Description]
|
|
188
|
-
Note: [Source4] reports a conflicting finding. See Contradictions section.
|
|
189
|
-
|
|
190
|
-
## Contradictions and Uncertainties
|
|
191
|
-
[Explicit list of unresolved disagreements]
|
|
192
|
-
|
|
193
|
-
## Gaps in Evidence
|
|
194
|
-
[What could not be determined from available sources]
|
|
195
|
-
|
|
196
|
-
## Source Registry
|
|
197
|
-
[Full list of all sources with grades and verification details]
|
|
198
|
-
```
|
|
199
|
-
|
|
200
|
-
## Best Practices
|
|
201
|
-
|
|
202
|
-
- Always search for disconfirming evidence, not just supporting evidence.
|
|
203
|
-
- Never assign HIGH confidence to a claim supported by only one source, regardless of its quality.
|
|
204
|
-
- Mark the date of your research prominently; findings may become outdated.
|
|
205
|
-
- Distinguish between "no evidence found" and "evidence of absence."
|
|
206
|
-
- When synthesizing, prefer systematic reviews and meta-analyses over individual studies.
|
|
207
|
-
- Keep the full source registry even if not all sources are cited in the final report.
|
|
208
|
-
|
|
209
|
-
## References
|
|
210
|
-
|
|
211
|
-
- Bates, M. J. (1989). The Design of Browsing and Berrypicking Techniques. *Online Review*, 13(5), 407-424.
|
|
212
|
-
- Booth, A. (2006). Clear and Present Questions: Formulating Questions for Evidence-Based Practice. *Library Hi Tech*, 24(3), 355-368.
|
|
213
|
-
- Grant, M. J. & Booth, A. (2009). A Typology of Reviews. *Health Information and Libraries Journal*, 26(2), 91-108.
|
|
@@ -1,204 +0,0 @@
|
|
|
1
|
-
---
|
|
2
|
-
name: deep-research-work
|
|
3
|
-
description: "Combine web search, content analysis, and source verification"
|
|
4
|
-
metadata:
|
|
5
|
-
openclaw:
|
|
6
|
-
emoji: "🌐"
|
|
7
|
-
category: "research"
|
|
8
|
-
subcategory: "deep-research"
|
|
9
|
-
keywords: ["web research", "content analysis", "source verification", "information extraction", "research workflow", "web scraping"]
|
|
10
|
-
source: "https://github.com/AcademicSkills/deep-research-work"
|
|
11
|
-
---
|
|
12
|
-
|
|
13
|
-
# Deep Research Work
|
|
14
|
-
|
|
15
|
-
A practical deep research workflow that combines web search, structured content analysis, and systematic source verification to produce comprehensive, trustworthy research outputs. Optimized for researchers who need to rapidly synthesize information from diverse online sources while maintaining academic standards of evidence quality.
|
|
16
|
-
|
|
17
|
-
## Overview
|
|
18
|
-
|
|
19
|
-
Academic research increasingly requires synthesizing information from beyond traditional journal databases: government datasets, technical documentation, industry reports, software repositories, news coverage, and expert commentary. Deep Research Work provides an operational workflow for conducting this kind of multi-source research efficiently while maintaining rigor. It covers search strategy formulation, content extraction, structured note-taking, source credibility assessment, and synthesis into coherent research narratives.
|
|
20
|
-
|
|
21
|
-
The workflow is designed to be completed in a single focused session (2-8 hours) and produces a structured research document with full source attribution. It is particularly useful for rapid literature scans, technology landscape assessments, policy research, and interdisciplinary investigations where no single database covers the full scope.
|
|
22
|
-
|
|
23
|
-
## Search Strategy Design
|
|
24
|
-
|
|
25
|
-
### Query Expansion Technique
|
|
26
|
-
|
|
27
|
-
```python
|
|
28
|
-
def generate_search_queries(topic: str, context: dict) -> list:
|
|
29
|
-
"""
|
|
30
|
-
Generate a comprehensive set of search queries using
|
|
31
|
-
systematic query expansion.
|
|
32
|
-
|
|
33
|
-
Expansion strategies:
|
|
34
|
-
1. Synonym expansion: use alternative terminology
|
|
35
|
-
2. Scope expansion: broaden/narrow the topic
|
|
36
|
-
3. Perspective expansion: different stakeholder views
|
|
37
|
-
4. Temporal expansion: historical and forward-looking
|
|
38
|
-
5. Geographic expansion: regional variations
|
|
39
|
-
"""
|
|
40
|
-
base_queries = [topic]
|
|
41
|
-
|
|
42
|
-
# Synonym expansion
|
|
43
|
-
synonyms = context.get('synonyms', [])
|
|
44
|
-
for syn in synonyms:
|
|
45
|
-
base_queries.append(syn)
|
|
46
|
-
|
|
47
|
-
# Scope expansion
|
|
48
|
-
broader = context.get('broader_topic', '')
|
|
49
|
-
narrower = context.get('sub_topics', [])
|
|
50
|
-
if broader:
|
|
51
|
-
base_queries.append(f"{broader} {topic}")
|
|
52
|
-
for sub in narrower:
|
|
53
|
-
base_queries.append(f"{topic} {sub}")
|
|
54
|
-
|
|
55
|
-
# Perspective expansion
|
|
56
|
-
perspectives = ['benefits', 'risks', 'challenges', 'future',
|
|
57
|
-
'criticism', 'comparison', 'case study']
|
|
58
|
-
for p in perspectives:
|
|
59
|
-
base_queries.append(f"{topic} {p}")
|
|
60
|
-
|
|
61
|
-
# Source-type targeting
|
|
62
|
-
source_types = ['systematic review', 'meta-analysis', 'white paper',
|
|
63
|
-
'technical report', 'dataset', 'open source']
|
|
64
|
-
for st in source_types:
|
|
65
|
-
base_queries.append(f"{topic} {st}")
|
|
66
|
-
|
|
67
|
-
return list(set(base_queries))
|
|
68
|
-
```
|
|
69
|
-
|
|
70
|
-
### Database Selection Matrix
|
|
71
|
-
|
|
72
|
-
| Source Type | Best Databases | When to Use |
|
|
73
|
-
|------------|---------------|-------------|
|
|
74
|
-
| Peer-reviewed articles | Google Scholar, Semantic Scholar, PubMed | Core academic evidence |
|
|
75
|
-
| Preprints | arXiv, bioRxiv, SSRN, medRxiv | Cutting-edge, pre-review findings |
|
|
76
|
-
| Government/institutional | Data.gov, WHO, OECD, national statistics | Official data, policy context |
|
|
77
|
-
| Technical documentation | GitHub, ReadTheDocs, official docs | Software, tools, methods |
|
|
78
|
-
| Industry reports | McKinsey, Gartner, CB Insights | Market context, trends |
|
|
79
|
-
| Patent databases | Google Patents, USPTO, Espacenet | Innovation landscape |
|
|
80
|
-
| News and media | Google News, specialized trade press | Current events, context |
|
|
81
|
-
|
|
82
|
-
## Content Extraction and Note-Taking
|
|
83
|
-
|
|
84
|
-
### Structured Extraction Template
|
|
85
|
-
|
|
86
|
-
For each source reviewed, extract the following:
|
|
87
|
-
|
|
88
|
-
```yaml
|
|
89
|
-
source_entry:
|
|
90
|
-
id: "S001"
|
|
91
|
-
url: "https://..."
|
|
92
|
-
title: "Title of the Source"
|
|
93
|
-
authors: ["Author A", "Author B"]
|
|
94
|
-
date: "2025-06"
|
|
95
|
-
type: "journal_article" # or preprint, report, blog, etc.
|
|
96
|
-
|
|
97
|
-
extraction:
|
|
98
|
-
main_claim: "One sentence summarizing the key claim or finding"
|
|
99
|
-
evidence_type: "empirical" # empirical, theoretical, anecdotal, opinion
|
|
100
|
-
methodology: "Randomized controlled trial, n=500"
|
|
101
|
-
key_data_points:
|
|
102
|
-
- "Finding 1: X increased by 23% (p < 0.01)"
|
|
103
|
-
- "Finding 2: No significant effect on Y"
|
|
104
|
-
limitations_noted: "Small sample from single institution"
|
|
105
|
-
relevant_quotes:
|
|
106
|
-
- page: 12
|
|
107
|
-
text: "Our results suggest that..."
|
|
108
|
-
|
|
109
|
-
assessment:
|
|
110
|
-
credibility: "high" # high, medium, low
|
|
111
|
-
relevance: "high" # high, medium, low
|
|
112
|
-
novelty: "medium" # high, medium, low
|
|
113
|
-
bias_concerns: "Funded by industry; potential conflict of interest"
|
|
114
|
-
```
|
|
115
|
-
|
|
116
|
-
### Progressive Summarization
|
|
117
|
-
|
|
118
|
-
Apply a layered note-taking approach:
|
|
119
|
-
|
|
120
|
-
1. **Layer 1 - Capture**: Save the full source with metadata (URL, date, authors).
|
|
121
|
-
2. **Layer 2 - Bold**: Highlight the most important passages (key findings, methods, conclusions).
|
|
122
|
-
3. **Layer 3 - Highlight**: From the bolded text, mark the essential takeaways for your research question.
|
|
123
|
-
4. **Layer 4 - Summary**: Write a 2-3 sentence summary in your own words.
|
|
124
|
-
5. **Layer 5 - Remix**: Connect the finding to your other sources and your research question.
|
|
125
|
-
|
|
126
|
-
## Source Verification Protocol
|
|
127
|
-
|
|
128
|
-
### Credibility Assessment Checklist
|
|
129
|
-
|
|
130
|
-
For each source, evaluate:
|
|
131
|
-
|
|
132
|
-
- [ ] **Authority**: Who is the author/organization? What are their credentials?
|
|
133
|
-
- [ ] **Accuracy**: Are claims supported by evidence? Can you verify the data?
|
|
134
|
-
- [ ] **Currency**: When was it published? Is the information still valid?
|
|
135
|
-
- [ ] **Coverage**: Does it address your question sufficiently?
|
|
136
|
-
- [ ] **Objectivity**: Is there apparent bias? Who funded the work?
|
|
137
|
-
- [ ] **Corroboration**: Do other independent sources support the same claims?
|
|
138
|
-
|
|
139
|
-
### Red Flags for Low-Quality Sources
|
|
140
|
-
|
|
141
|
-
| Red Flag | Action |
|
|
142
|
-
|----------|--------|
|
|
143
|
-
| No author attribution | Downgrade credibility; seek alternative source |
|
|
144
|
-
| No date published | Treat as potentially outdated |
|
|
145
|
-
| Extraordinary claims without evidence | Require independent corroboration |
|
|
146
|
-
| Known predatory journal | Exclude from primary evidence |
|
|
147
|
-
| Single anonymous blog post | Use only as lead to find primary sources |
|
|
148
|
-
| Circular citations | Trace back to the original source |
|
|
149
|
-
|
|
150
|
-
## Synthesis Workflow
|
|
151
|
-
|
|
152
|
-
### From Notes to Narrative
|
|
153
|
-
|
|
154
|
-
```
|
|
155
|
-
Step 1: Cluster
|
|
156
|
-
Group extracted notes by theme or sub-question.
|
|
157
|
-
Use tags from your extraction template.
|
|
158
|
-
|
|
159
|
-
Step 2: Compare
|
|
160
|
-
Within each cluster, compare findings across sources.
|
|
161
|
-
Note agreements, contradictions, and gaps.
|
|
162
|
-
|
|
163
|
-
Step 3: Evaluate
|
|
164
|
-
Weight evidence by source credibility and recency.
|
|
165
|
-
Higher-quality sources take precedence when sources conflict.
|
|
166
|
-
|
|
167
|
-
Step 4: Narrate
|
|
168
|
-
Write a synthesis paragraph for each cluster that:
|
|
169
|
-
- States the overall finding
|
|
170
|
-
- Cites the supporting sources
|
|
171
|
-
- Notes any caveats or contradictions
|
|
172
|
-
- Identifies remaining gaps
|
|
173
|
-
|
|
174
|
-
Step 5: Integrate
|
|
175
|
-
Connect clusters into a coherent narrative.
|
|
176
|
-
Highlight cross-cutting themes and implications.
|
|
177
|
-
```
|
|
178
|
-
|
|
179
|
-
### Output Quality Checklist
|
|
180
|
-
|
|
181
|
-
Before finalizing your research output:
|
|
182
|
-
|
|
183
|
-
- [ ] Every factual claim has at least one source citation
|
|
184
|
-
- [ ] Contradictory evidence is explicitly acknowledged
|
|
185
|
-
- [ ] Source quality is visible (not all sources treated equally)
|
|
186
|
-
- [ ] Gaps in knowledge are clearly identified
|
|
187
|
-
- [ ] The search methodology is documented for reproducibility
|
|
188
|
-
- [ ] Dates of all searches are recorded
|
|
189
|
-
- [ ] The output answers the original research question
|
|
190
|
-
|
|
191
|
-
## Best Practices
|
|
192
|
-
|
|
193
|
-
- Set a time limit before starting. Research can expand indefinitely without constraints.
|
|
194
|
-
- Use a reference manager (Zotero, Mendeley) from the start, even for informal research.
|
|
195
|
-
- Save web pages as PDF or archive snapshots (Wayback Machine) to prevent link rot.
|
|
196
|
-
- Distinguish between primary sources (original data/study) and secondary sources (reporting on the study).
|
|
197
|
-
- When a source cites a finding, always try to trace back to the original source.
|
|
198
|
-
- Document negative results: sources searched that did not yield relevant information.
|
|
199
|
-
|
|
200
|
-
## References
|
|
201
|
-
|
|
202
|
-
- Booth, A., Sutton, A., & Papaioannou, D. (2016). *Systematic Approaches to a Successful Literature Review* (2nd ed.). Sage.
|
|
203
|
-
- Forte, T. (2022). *Building a Second Brain*. Atria Books.
|
|
204
|
-
- Machi, L. A. & McEvoy, B. T. (2016). *The Literature Review* (3rd ed.). Corwin Press.
|