@wentorai/research-plugins 1.2.2 → 1.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (141) hide show
  1. package/README.md +16 -8
  2. package/openclaw.plugin.json +10 -3
  3. package/package.json +2 -5
  4. package/skills/analysis/dataviz/SKILL.md +25 -0
  5. package/skills/analysis/dataviz/chart-image-generator/SKILL.md +1 -1
  6. package/skills/analysis/econometrics/SKILL.md +23 -0
  7. package/skills/analysis/econometrics/robustness-checks/SKILL.md +1 -1
  8. package/skills/analysis/statistics/SKILL.md +21 -0
  9. package/skills/analysis/statistics/data-anomaly-detection/SKILL.md +1 -1
  10. package/skills/analysis/statistics/ml-experiment-tracker/SKILL.md +1 -1
  11. package/skills/analysis/statistics/{senior-data-scientist-guide → modeling-strategy-guide}/SKILL.md +5 -5
  12. package/skills/analysis/wrangling/SKILL.md +21 -0
  13. package/skills/analysis/wrangling/csv-data-analyzer/SKILL.md +1 -1
  14. package/skills/analysis/wrangling/data-cog-guide/SKILL.md +1 -1
  15. package/skills/domains/ai-ml/SKILL.md +37 -0
  16. package/skills/domains/biomedical/SKILL.md +28 -0
  17. package/skills/domains/biomedical/genomas-guide/SKILL.md +1 -1
  18. package/skills/domains/biomedical/med-researcher-guide/SKILL.md +1 -1
  19. package/skills/domains/biomedical/medgeclaw-guide/SKILL.md +1 -1
  20. package/skills/domains/business/SKILL.md +17 -0
  21. package/skills/domains/business/architecture-design-guide/SKILL.md +1 -1
  22. package/skills/domains/chemistry/SKILL.md +19 -0
  23. package/skills/domains/chemistry/computational-chemistry-guide/SKILL.md +1 -1
  24. package/skills/domains/cs/SKILL.md +21 -0
  25. package/skills/domains/ecology/SKILL.md +16 -0
  26. package/skills/domains/economics/SKILL.md +20 -0
  27. package/skills/domains/economics/post-labor-economics/SKILL.md +1 -1
  28. package/skills/domains/economics/pricing-psychology-guide/SKILL.md +1 -1
  29. package/skills/domains/education/SKILL.md +19 -0
  30. package/skills/domains/education/academic-study-methods/SKILL.md +1 -1
  31. package/skills/domains/education/edumcp-guide/SKILL.md +1 -1
  32. package/skills/domains/finance/SKILL.md +19 -0
  33. package/skills/domains/finance/akshare-finance-data/SKILL.md +1 -1
  34. package/skills/domains/finance/options-analytics-agent-guide/SKILL.md +1 -1
  35. package/skills/domains/finance/stata-accounting-research/SKILL.md +1 -1
  36. package/skills/domains/geoscience/SKILL.md +17 -0
  37. package/skills/domains/humanities/SKILL.md +16 -0
  38. package/skills/domains/humanities/history-research-guide/SKILL.md +1 -1
  39. package/skills/domains/humanities/political-history-guide/SKILL.md +1 -1
  40. package/skills/domains/law/SKILL.md +19 -0
  41. package/skills/domains/math/SKILL.md +17 -0
  42. package/skills/domains/pharma/SKILL.md +17 -0
  43. package/skills/domains/physics/SKILL.md +16 -0
  44. package/skills/domains/social-science/SKILL.md +17 -0
  45. package/skills/domains/social-science/sociology-research-methods/SKILL.md +1 -1
  46. package/skills/literature/discovery/SKILL.md +20 -0
  47. package/skills/literature/discovery/paper-recommendation-guide/SKILL.md +1 -1
  48. package/skills/literature/discovery/semantic-paper-radar/SKILL.md +1 -1
  49. package/skills/literature/fulltext/SKILL.md +26 -0
  50. package/skills/literature/metadata/SKILL.md +35 -0
  51. package/skills/literature/metadata/doi-content-negotiation/SKILL.md +4 -0
  52. package/skills/literature/metadata/doi-resolution-guide/SKILL.md +4 -0
  53. package/skills/literature/metadata/orcid-api/SKILL.md +4 -0
  54. package/skills/literature/metadata/orcid-integration-guide/SKILL.md +4 -0
  55. package/skills/literature/search/SKILL.md +43 -0
  56. package/skills/literature/search/paper-search-mcp-guide/SKILL.md +1 -1
  57. package/skills/research/automation/SKILL.md +21 -0
  58. package/skills/research/deep-research/SKILL.md +24 -0
  59. package/skills/research/deep-research/auto-deep-research-guide/SKILL.md +1 -1
  60. package/skills/research/deep-research/in-depth-research-guide/SKILL.md +1 -1
  61. package/skills/research/funding/SKILL.md +20 -0
  62. package/skills/research/methodology/SKILL.md +24 -0
  63. package/skills/research/paper-review/SKILL.md +19 -0
  64. package/skills/research/paper-review/paper-critique-framework/SKILL.md +1 -1
  65. package/skills/tools/code-exec/SKILL.md +18 -0
  66. package/skills/tools/diagram/SKILL.md +20 -0
  67. package/skills/tools/document/SKILL.md +21 -0
  68. package/skills/tools/knowledge-graph/SKILL.md +21 -0
  69. package/skills/tools/ocr-translate/SKILL.md +18 -0
  70. package/skills/tools/ocr-translate/handwriting-recognition-guide/SKILL.md +2 -0
  71. package/skills/tools/ocr-translate/latex-ocr-guide/SKILL.md +2 -0
  72. package/skills/tools/scraping/SKILL.md +17 -0
  73. package/skills/writing/citation/SKILL.md +33 -0
  74. package/skills/writing/citation/zotfile-attachment-guide/SKILL.md +2 -0
  75. package/skills/writing/composition/SKILL.md +22 -0
  76. package/skills/writing/composition/research-paper-writer/SKILL.md +1 -1
  77. package/skills/writing/composition/scientific-writing-wrapper/SKILL.md +1 -1
  78. package/skills/writing/latex/SKILL.md +22 -0
  79. package/skills/writing/latex/academic-writing-latex/SKILL.md +1 -1
  80. package/skills/writing/latex/latex-drawing-guide/SKILL.md +1 -1
  81. package/skills/writing/polish/SKILL.md +20 -0
  82. package/skills/writing/polish/chinese-text-humanizer/SKILL.md +1 -1
  83. package/skills/writing/templates/SKILL.md +22 -0
  84. package/skills/writing/templates/beamer-presentation-guide/SKILL.md +1 -1
  85. package/skills/writing/templates/scientific-article-pdf/SKILL.md +1 -1
  86. package/skills/analysis/dataviz/citation-map-guide/SKILL.md +0 -184
  87. package/skills/analysis/dataviz/data-visualization-principles/SKILL.md +0 -171
  88. package/skills/analysis/econometrics/empirical-paper-analysis/SKILL.md +0 -192
  89. package/skills/analysis/econometrics/panel-data-regression-workflow/SKILL.md +0 -267
  90. package/skills/analysis/econometrics/stata-regression/SKILL.md +0 -117
  91. package/skills/analysis/statistics/general-statistics-guide/SKILL.md +0 -226
  92. package/skills/analysis/statistics/infiagent-benchmark-guide/SKILL.md +0 -106
  93. package/skills/analysis/statistics/pywayne-statistics-guide/SKILL.md +0 -192
  94. package/skills/analysis/statistics/quantitative-methods-guide/SKILL.md +0 -193
  95. package/skills/analysis/wrangling/claude-data-analysis-guide/SKILL.md +0 -100
  96. package/skills/analysis/wrangling/open-data-scientist-guide/SKILL.md +0 -197
  97. package/skills/domains/ai-ml/annotated-dl-papers-guide/SKILL.md +0 -159
  98. package/skills/domains/humanities/digital-humanities-methods/SKILL.md +0 -232
  99. package/skills/domains/law/legal-research-methods/SKILL.md +0 -190
  100. package/skills/domains/social-science/sociology-research-guide/SKILL.md +0 -238
  101. package/skills/literature/discovery/arxiv-paper-monitoring/SKILL.md +0 -233
  102. package/skills/literature/discovery/paper-tracking-guide/SKILL.md +0 -211
  103. package/skills/literature/fulltext/zotero-scihub-guide/SKILL.md +0 -168
  104. package/skills/literature/search/arxiv-osiris/SKILL.md +0 -199
  105. package/skills/literature/search/deepgit-search-guide/SKILL.md +0 -147
  106. package/skills/literature/search/multi-database-literature-search/SKILL.md +0 -198
  107. package/skills/literature/search/papers-chat-guide/SKILL.md +0 -194
  108. package/skills/literature/search/pasa-paper-search-guide/SKILL.md +0 -138
  109. package/skills/literature/search/scientify-literature-survey/SKILL.md +0 -203
  110. package/skills/research/automation/ai-scientist-guide/SKILL.md +0 -228
  111. package/skills/research/automation/coexist-ai-guide/SKILL.md +0 -149
  112. package/skills/research/automation/foam-agent-guide/SKILL.md +0 -203
  113. package/skills/research/automation/research-paper-orchestrator/SKILL.md +0 -254
  114. package/skills/research/deep-research/academic-deep-research/SKILL.md +0 -190
  115. package/skills/research/deep-research/cognitive-kernel-guide/SKILL.md +0 -200
  116. package/skills/research/deep-research/corvus-research-guide/SKILL.md +0 -132
  117. package/skills/research/deep-research/deep-research-pro/SKILL.md +0 -213
  118. package/skills/research/deep-research/deep-research-work/SKILL.md +0 -204
  119. package/skills/research/deep-research/research-cog/SKILL.md +0 -153
  120. package/skills/research/methodology/academic-mentor-guide/SKILL.md +0 -169
  121. package/skills/research/methodology/deep-innovator-guide/SKILL.md +0 -242
  122. package/skills/research/methodology/research-pipeline-units-guide/SKILL.md +0 -169
  123. package/skills/research/paper-review/paper-compare-guide/SKILL.md +0 -238
  124. package/skills/research/paper-review/paper-digest-guide/SKILL.md +0 -240
  125. package/skills/research/paper-review/paper-research-assistant/SKILL.md +0 -231
  126. package/skills/research/paper-review/research-quality-filter/SKILL.md +0 -261
  127. package/skills/tools/code-exec/contextplus-mcp-guide/SKILL.md +0 -110
  128. package/skills/tools/diagram/clawphd-guide/SKILL.md +0 -149
  129. package/skills/tools/diagram/scientific-graphical-abstract/SKILL.md +0 -201
  130. package/skills/tools/document/md2pdf-xelatex/SKILL.md +0 -212
  131. package/skills/tools/document/openpaper-guide/SKILL.md +0 -232
  132. package/skills/tools/document/weknora-guide/SKILL.md +0 -216
  133. package/skills/tools/knowledge-graph/mimir-memory-guide/SKILL.md +0 -135
  134. package/skills/tools/knowledge-graph/open-webui-tools-guide/SKILL.md +0 -156
  135. package/skills/tools/ocr-translate/formula-recognition-guide/SKILL.md +0 -367
  136. package/skills/tools/ocr-translate/math-equation-renderer/SKILL.md +0 -198
  137. package/skills/tools/scraping/api-data-collection-guide/SKILL.md +0 -301
  138. package/skills/writing/citation/academic-citation-manager-guide/SKILL.md +0 -182
  139. package/skills/writing/composition/opendraft-thesis-guide/SKILL.md +0 -200
  140. package/skills/writing/composition/paper-debugger-guide/SKILL.md +0 -143
  141. package/skills/writing/composition/paperforge-guide/SKILL.md +0 -205
@@ -1,132 +0,0 @@
1
- ---
2
- name: corvus-research-guide
3
- description: "Multi-agent AI research with semantic search and citation snowballing"
4
- metadata:
5
- openclaw:
6
- emoji: "🐦‍⬛"
7
- category: "research"
8
- subcategory: "deep-research"
9
- keywords: ["Corvus", "multi-agent", "semantic search", "citation snowballing", "research synthesis", "AI research"]
10
- source: "https://github.com/corvus-research/corvus"
11
- ---
12
-
13
- # Corvus Research Guide
14
-
15
- ## Overview
16
-
17
- Corvus is a multi-agent AI research system that combines semantic search, forward/backward citation snowballing, and synthesis to conduct thorough literature investigations. It iteratively expands search results by following citation chains, identifies research gaps, and generates structured research briefs with full provenance.
18
-
19
- ## Architecture
20
-
21
- ### Agent Pipeline
22
-
23
- ```
24
- Query → Semantic Search Agent
25
-
26
- Citation Snowball Agent (forward + backward)
27
-
28
- Relevance Filter Agent
29
-
30
- Synthesis Agent
31
-
32
- Report with Citation Graph
33
- ```
34
-
35
- ### Key Features
36
-
37
- 1. **Semantic search**: Uses embedding-based search across Semantic Scholar, OpenAlex
38
- 2. **Citation snowballing**: Iteratively follows references (backward) and citations (forward) to discover related work
39
- 3. **Relevance scoring**: AI-based relevance assessment at each expansion step
40
- 4. **Provenance tracking**: Every claim linked to source papers
41
- 5. **Gap identification**: Identifies under-explored research areas
42
-
43
- ## Usage
44
-
45
- ```python
46
- from corvus import ResearchAgent
47
-
48
- agent = ResearchAgent(
49
- llm_provider="anthropic",
50
- search_backends=["semantic_scholar", "openalex"],
51
- max_snowball_depth=2,
52
- )
53
-
54
- # Conduct deep research
55
- result = agent.research(
56
- query="What are the current approaches to continual learning "
57
- "in large language models?",
58
- initial_papers=20,
59
- snowball_per_paper=5,
60
- )
61
-
62
- # Access results
63
- print(f"Papers found: {len(result.papers)}")
64
- print(f"Unique clusters: {len(result.clusters)}")
65
- print(f"\nSynthesis:\n{result.synthesis}")
66
-
67
- # Export citation graph
68
- result.export_graph("citation_network.gexf")
69
-
70
- # Export bibliography
71
- result.export_bibtex("references.bib")
72
- ```
73
-
74
- ### Snowballing Configuration
75
-
76
- ```python
77
- agent = ResearchAgent(
78
- snowball_config={
79
- "max_depth": 3, # Citation chain depth
80
- "backward_limit": 10, # References per paper
81
- "forward_limit": 10, # Citations per paper
82
- "relevance_threshold": 0.7, # Min relevance to continue
83
- "year_filter": 2020, # Only papers from 2020+
84
- }
85
- )
86
- ```
87
-
88
- ### Research Modes
89
-
90
- ```python
91
- # Broad survey mode
92
- result = agent.research(query, mode="survey",
93
- initial_papers=50)
94
-
95
- # Focused deep-dive
96
- result = agent.research(query, mode="focused",
97
- initial_papers=10,
98
- snowball_depth=3)
99
-
100
- # Gap analysis
101
- result = agent.research(query, mode="gap_analysis")
102
- # Returns underexplored subtopics and suggested directions
103
- ```
104
-
105
- ## Output Format
106
-
107
- ```python
108
- # Structured research brief
109
- brief = result.generate_brief()
110
-
111
- # Contains:
112
- # - Research question
113
- # - Methodology (search strategy, databases, snowball depth)
114
- # - Key themes (clustered by topic)
115
- # - Timeline (research evolution over time)
116
- # - Gap analysis (underexplored areas)
117
- # - Bibliography (all papers with citation counts)
118
-
119
- brief.save("research_brief.md")
120
- ```
121
-
122
- ## Use Cases
123
-
124
- 1. **Literature reviews**: Comprehensive coverage via snowballing
125
- 2. **Research gap identification**: Find underexplored subtopics
126
- 3. **Trend analysis**: Track research evolution through citation chains
127
- 4. **Grant proposals**: Quick evidence of research need
128
-
129
- ## References
130
-
131
- - [Corvus GitHub](https://github.com/corvus-research/corvus)
132
- - Wohlin, C. (2014). "Guidelines for snowballing in systematic literature studies." *EASE 2014*.
@@ -1,213 +0,0 @@
1
- ---
2
- name: deep-research-pro
3
- description: "Multi-source deep research agent with verification and synthesis"
4
- metadata:
5
- openclaw:
6
- emoji: "🕵️"
7
- category: "research"
8
- subcategory: "deep-research"
9
- keywords: ["deep research", "multi-source", "source verification", "research agent", "evidence synthesis", "fact checking"]
10
- source: "https://github.com/AcademicSkills/deep-research-pro"
11
- ---
12
-
13
- # Deep Research Pro
14
-
15
- A professional-grade deep research methodology that coordinates multi-source information gathering, cross-reference verification, and structured synthesis. Designed for research tasks that require high confidence in factual accuracy, comprehensive coverage, and traceable evidence chains.
16
-
17
- ## Overview
18
-
19
- Deep Research Pro implements an agent-based methodology where the research process is decomposed into specialized phases: query decomposition, parallel source gathering, cross-reference verification, contradiction resolution, and structured synthesis. Unlike simple search-and-summarize approaches, this skill emphasizes source triangulation, evidence grading, and explicit uncertainty marking.
20
-
21
- The methodology is particularly valuable for literature reviews, technology assessments, policy analyses, and any research task where decision-makers need to trust the completeness and accuracy of the findings. Every claim in the final output is linked to at least one verified source, with confidence levels assigned based on the quality and agreement of the evidence.
22
-
23
- ## Research Agent Architecture
24
-
25
- ### Phase 1: Query Decomposition
26
-
27
- ```python
28
- def decompose_research_query(query: str) -> dict:
29
- """
30
- Break a complex research question into atomic sub-questions
31
- that can be independently investigated.
32
-
33
- Example:
34
- Input: "What is the current state of quantum computing for
35
- drug discovery, and which companies are leading?"
36
-
37
- Output: {
38
- 'sub_questions': [
39
- 'What quantum computing methods are used in drug discovery?',
40
- 'What are the computational advantages over classical methods?',
41
- 'Which companies are actively working on quantum drug discovery?',
42
- 'What drugs or molecules have been studied using quantum methods?',
43
- 'What are the current limitations and timeline to practical use?'
44
- ],
45
- 'cross_cutting_themes': [
46
- 'Cost-benefit vs classical computing',
47
- 'Regulatory considerations',
48
- 'Academic vs industry progress'
49
- ],
50
- 'source_strategy': {
51
- 'academic': ['PubMed', 'arXiv', 'Google Scholar'],
52
- 'industry': ['company websites', 'press releases', 'patent databases'],
53
- 'grey': ['government reports', 'consulting firm analyses']
54
- }
55
- }
56
- """
57
- pass # Implemented by the agent's reasoning
58
- ```
59
-
60
- ### Phase 2: Parallel Source Gathering
61
-
62
- For each sub-question, gather evidence from multiple independent source types:
63
-
64
- | Source Type | Examples | Strengths | Limitations |
65
- |------------|---------|-----------|-------------|
66
- | **Peer-reviewed** | Journal articles, conference papers | Rigorous review | Publication lag |
67
- | **Preprints** | arXiv, bioRxiv, SSRN | Current | Not peer-reviewed |
68
- | **Institutional** | WHO, NIH, government agencies | Authoritative | May be conservative |
69
- | **Industry** | Company blogs, press releases | Current, practical | Biased toward own products |
70
- | **News/Media** | Science journalism, trade publications | Accessible | May oversimplify |
71
- | **Expert opinion** | Interviews, blog posts, talks | Nuanced | Subjective |
72
-
73
- ### Phase 3: Source Verification
74
-
75
- ```python
76
- def verify_source(source: dict) -> dict:
77
- """
78
- Assess the credibility and reliability of a single source.
79
- """
80
- verification = {
81
- 'source_id': source['id'],
82
- 'url': source['url'],
83
- 'checks': {}
84
- }
85
-
86
- # Check 1: Author credibility
87
- verification['checks']['author'] = {
88
- 'identifiable': bool(source.get('author')),
89
- 'affiliated': bool(source.get('institution')),
90
- 'h_index_available': bool(source.get('h_index')),
91
- 'domain_expert': source.get('expertise_match', False)
92
- }
93
-
94
- # Check 2: Publication venue
95
- verification['checks']['venue'] = {
96
- 'peer_reviewed': source.get('peer_reviewed', False),
97
- 'impact_factor': source.get('impact_factor'),
98
- 'known_predatory': source.get('predatory_flag', False)
99
- }
100
-
101
- # Check 3: Currency
102
- verification['checks']['currency'] = {
103
- 'publication_year': source.get('year'),
104
- 'is_recent': source.get('year', 0) >= 2023,
105
- 'superseded': source.get('retracted', False)
106
- }
107
-
108
- # Check 4: Corroboration
109
- verification['checks']['corroboration'] = {
110
- 'cited_by_count': source.get('citation_count', 0),
111
- 'independent_confirmation': source.get('replicated', False),
112
- 'consistent_with_consensus': source.get('consensus_aligned', None)
113
- }
114
-
115
- # Overall confidence score
116
- score = sum([
117
- verification['checks']['author']['identifiable'],
118
- verification['checks']['venue']['peer_reviewed'],
119
- not verification['checks']['venue']['known_predatory'],
120
- verification['checks']['currency']['is_recent'],
121
- verification['checks']['corroboration']['cited_by_count'] > 5
122
- ]) / 5.0
123
-
124
- verification['confidence'] = round(score, 2)
125
- verification['grade'] = (
126
- 'A' if score >= 0.8 else
127
- 'B' if score >= 0.6 else
128
- 'C' if score >= 0.4 else 'D'
129
- )
130
- return verification
131
- ```
132
-
133
- ## Cross-Reference and Contradiction Resolution
134
-
135
- ### Triangulation Protocol
136
-
137
- For each factual claim in the research output:
138
-
139
- 1. **Identify the claim**: Extract a specific, testable statement.
140
- 2. **Find supporting sources**: At least 2 independent sources.
141
- 3. **Check for contradicting sources**: Actively search for disconfirming evidence.
142
- 4. **Assess source quality**: Weight evidence by source grade (A > B > C > D).
143
- 5. **Assign confidence level**: Based on evidence agreement and quality.
144
-
145
- ```
146
- Confidence Levels:
147
- HIGH: 3+ Grade A/B sources agree, no contradictions
148
- MEDIUM: 2+ sources agree, minor contradictions or lower-quality sources
149
- LOW: Single source, or significant contradictions among sources
150
- UNCERTAIN: Conflicting evidence of similar quality; state both positions
151
- ```
152
-
153
- ### Contradiction Handling
154
-
155
- When sources disagree:
156
-
157
- 1. Check if the disagreement is due to different definitions, timeframes, or contexts.
158
- 2. Check if one source has been superseded or corrected.
159
- 3. If genuine disagreement exists, present both positions with their evidence.
160
- 4. Note the disagreement explicitly: "Source A (Grade B, 2024) reports X, while Source B (Grade A, 2023) reports Y. The discrepancy may be due to [possible explanation]."
161
-
162
- ## Structured Synthesis Output
163
-
164
- ### Report Template
165
-
166
- ```markdown
167
- # Research Report: [Topic]
168
-
169
- ## Executive Summary
170
- [3-5 sentence overview of key findings with confidence levels]
171
-
172
- ## Methodology
173
- - Sub-questions investigated: [list]
174
- - Sources consulted: [count by type]
175
- - Date range of sources: [range]
176
- - Verification standard: [triangulation with Grade B+ sources]
177
-
178
- ## Findings
179
-
180
- ### Finding 1: [Title]
181
- **Confidence: HIGH**
182
- [Description with inline source references]
183
- Sources: [Source1 (Grade A)], [Source2 (Grade B)], [Source3 (Grade A)]
184
-
185
- ### Finding 2: [Title]
186
- **Confidence: MEDIUM**
187
- [Description]
188
- Note: [Source4] reports a conflicting finding. See Contradictions section.
189
-
190
- ## Contradictions and Uncertainties
191
- [Explicit list of unresolved disagreements]
192
-
193
- ## Gaps in Evidence
194
- [What could not be determined from available sources]
195
-
196
- ## Source Registry
197
- [Full list of all sources with grades and verification details]
198
- ```
199
-
200
- ## Best Practices
201
-
202
- - Always search for disconfirming evidence, not just supporting evidence.
203
- - Never assign HIGH confidence to a claim supported by only one source, regardless of its quality.
204
- - Mark the date of your research prominently; findings may become outdated.
205
- - Distinguish between "no evidence found" and "evidence of absence."
206
- - When synthesizing, prefer systematic reviews and meta-analyses over individual studies.
207
- - Keep the full source registry even if not all sources are cited in the final report.
208
-
209
- ## References
210
-
211
- - Bates, M. J. (1989). The Design of Browsing and Berrypicking Techniques. *Online Review*, 13(5), 407-424.
212
- - Booth, A. (2006). Clear and Present Questions: Formulating Questions for Evidence-Based Practice. *Library Hi Tech*, 24(3), 355-368.
213
- - Grant, M. J. & Booth, A. (2009). A Typology of Reviews. *Health Information and Libraries Journal*, 26(2), 91-108.
@@ -1,204 +0,0 @@
1
- ---
2
- name: deep-research-work
3
- description: "Combine web search, content analysis, and source verification"
4
- metadata:
5
- openclaw:
6
- emoji: "🌐"
7
- category: "research"
8
- subcategory: "deep-research"
9
- keywords: ["web research", "content analysis", "source verification", "information extraction", "research workflow", "web scraping"]
10
- source: "https://github.com/AcademicSkills/deep-research-work"
11
- ---
12
-
13
- # Deep Research Work
14
-
15
- A practical deep research workflow that combines web search, structured content analysis, and systematic source verification to produce comprehensive, trustworthy research outputs. Optimized for researchers who need to rapidly synthesize information from diverse online sources while maintaining academic standards of evidence quality.
16
-
17
- ## Overview
18
-
19
- Academic research increasingly requires synthesizing information from beyond traditional journal databases: government datasets, technical documentation, industry reports, software repositories, news coverage, and expert commentary. Deep Research Work provides an operational workflow for conducting this kind of multi-source research efficiently while maintaining rigor. It covers search strategy formulation, content extraction, structured note-taking, source credibility assessment, and synthesis into coherent research narratives.
20
-
21
- The workflow is designed to be completed in a single focused session (2-8 hours) and produces a structured research document with full source attribution. It is particularly useful for rapid literature scans, technology landscape assessments, policy research, and interdisciplinary investigations where no single database covers the full scope.
22
-
23
- ## Search Strategy Design
24
-
25
- ### Query Expansion Technique
26
-
27
- ```python
28
- def generate_search_queries(topic: str, context: dict) -> list:
29
- """
30
- Generate a comprehensive set of search queries using
31
- systematic query expansion.
32
-
33
- Expansion strategies:
34
- 1. Synonym expansion: use alternative terminology
35
- 2. Scope expansion: broaden/narrow the topic
36
- 3. Perspective expansion: different stakeholder views
37
- 4. Temporal expansion: historical and forward-looking
38
- 5. Geographic expansion: regional variations
39
- """
40
- base_queries = [topic]
41
-
42
- # Synonym expansion
43
- synonyms = context.get('synonyms', [])
44
- for syn in synonyms:
45
- base_queries.append(syn)
46
-
47
- # Scope expansion
48
- broader = context.get('broader_topic', '')
49
- narrower = context.get('sub_topics', [])
50
- if broader:
51
- base_queries.append(f"{broader} {topic}")
52
- for sub in narrower:
53
- base_queries.append(f"{topic} {sub}")
54
-
55
- # Perspective expansion
56
- perspectives = ['benefits', 'risks', 'challenges', 'future',
57
- 'criticism', 'comparison', 'case study']
58
- for p in perspectives:
59
- base_queries.append(f"{topic} {p}")
60
-
61
- # Source-type targeting
62
- source_types = ['systematic review', 'meta-analysis', 'white paper',
63
- 'technical report', 'dataset', 'open source']
64
- for st in source_types:
65
- base_queries.append(f"{topic} {st}")
66
-
67
- return list(set(base_queries))
68
- ```
69
-
70
- ### Database Selection Matrix
71
-
72
- | Source Type | Best Databases | When to Use |
73
- |------------|---------------|-------------|
74
- | Peer-reviewed articles | Google Scholar, Semantic Scholar, PubMed | Core academic evidence |
75
- | Preprints | arXiv, bioRxiv, SSRN, medRxiv | Cutting-edge, pre-review findings |
76
- | Government/institutional | Data.gov, WHO, OECD, national statistics | Official data, policy context |
77
- | Technical documentation | GitHub, ReadTheDocs, official docs | Software, tools, methods |
78
- | Industry reports | McKinsey, Gartner, CB Insights | Market context, trends |
79
- | Patent databases | Google Patents, USPTO, Espacenet | Innovation landscape |
80
- | News and media | Google News, specialized trade press | Current events, context |
81
-
82
- ## Content Extraction and Note-Taking
83
-
84
- ### Structured Extraction Template
85
-
86
- For each source reviewed, extract the following:
87
-
88
- ```yaml
89
- source_entry:
90
- id: "S001"
91
- url: "https://..."
92
- title: "Title of the Source"
93
- authors: ["Author A", "Author B"]
94
- date: "2025-06"
95
- type: "journal_article" # or preprint, report, blog, etc.
96
-
97
- extraction:
98
- main_claim: "One sentence summarizing the key claim or finding"
99
- evidence_type: "empirical" # empirical, theoretical, anecdotal, opinion
100
- methodology: "Randomized controlled trial, n=500"
101
- key_data_points:
102
- - "Finding 1: X increased by 23% (p < 0.01)"
103
- - "Finding 2: No significant effect on Y"
104
- limitations_noted: "Small sample from single institution"
105
- relevant_quotes:
106
- - page: 12
107
- text: "Our results suggest that..."
108
-
109
- assessment:
110
- credibility: "high" # high, medium, low
111
- relevance: "high" # high, medium, low
112
- novelty: "medium" # high, medium, low
113
- bias_concerns: "Funded by industry; potential conflict of interest"
114
- ```
115
-
116
- ### Progressive Summarization
117
-
118
- Apply a layered note-taking approach:
119
-
120
- 1. **Layer 1 - Capture**: Save the full source with metadata (URL, date, authors).
121
- 2. **Layer 2 - Bold**: Highlight the most important passages (key findings, methods, conclusions).
122
- 3. **Layer 3 - Highlight**: From the bolded text, mark the essential takeaways for your research question.
123
- 4. **Layer 4 - Summary**: Write a 2-3 sentence summary in your own words.
124
- 5. **Layer 5 - Remix**: Connect the finding to your other sources and your research question.
125
-
126
- ## Source Verification Protocol
127
-
128
- ### Credibility Assessment Checklist
129
-
130
- For each source, evaluate:
131
-
132
- - [ ] **Authority**: Who is the author/organization? What are their credentials?
133
- - [ ] **Accuracy**: Are claims supported by evidence? Can you verify the data?
134
- - [ ] **Currency**: When was it published? Is the information still valid?
135
- - [ ] **Coverage**: Does it address your question sufficiently?
136
- - [ ] **Objectivity**: Is there apparent bias? Who funded the work?
137
- - [ ] **Corroboration**: Do other independent sources support the same claims?
138
-
139
- ### Red Flags for Low-Quality Sources
140
-
141
- | Red Flag | Action |
142
- |----------|--------|
143
- | No author attribution | Downgrade credibility; seek alternative source |
144
- | No date published | Treat as potentially outdated |
145
- | Extraordinary claims without evidence | Require independent corroboration |
146
- | Known predatory journal | Exclude from primary evidence |
147
- | Single anonymous blog post | Use only as lead to find primary sources |
148
- | Circular citations | Trace back to the original source |
149
-
150
- ## Synthesis Workflow
151
-
152
- ### From Notes to Narrative
153
-
154
- ```
155
- Step 1: Cluster
156
- Group extracted notes by theme or sub-question.
157
- Use tags from your extraction template.
158
-
159
- Step 2: Compare
160
- Within each cluster, compare findings across sources.
161
- Note agreements, contradictions, and gaps.
162
-
163
- Step 3: Evaluate
164
- Weight evidence by source credibility and recency.
165
- Higher-quality sources take precedence when sources conflict.
166
-
167
- Step 4: Narrate
168
- Write a synthesis paragraph for each cluster that:
169
- - States the overall finding
170
- - Cites the supporting sources
171
- - Notes any caveats or contradictions
172
- - Identifies remaining gaps
173
-
174
- Step 5: Integrate
175
- Connect clusters into a coherent narrative.
176
- Highlight cross-cutting themes and implications.
177
- ```
178
-
179
- ### Output Quality Checklist
180
-
181
- Before finalizing your research output:
182
-
183
- - [ ] Every factual claim has at least one source citation
184
- - [ ] Contradictory evidence is explicitly acknowledged
185
- - [ ] Source quality is visible (not all sources treated equally)
186
- - [ ] Gaps in knowledge are clearly identified
187
- - [ ] The search methodology is documented for reproducibility
188
- - [ ] Dates of all searches are recorded
189
- - [ ] The output answers the original research question
190
-
191
- ## Best Practices
192
-
193
- - Set a time limit before starting. Research can expand indefinitely without constraints.
194
- - Use a reference manager (Zotero, Mendeley) from the start, even for informal research.
195
- - Save web pages as PDF or archive snapshots (Wayback Machine) to prevent link rot.
196
- - Distinguish between primary sources (original data/study) and secondary sources (reporting on the study).
197
- - When a source cites a finding, always try to trace back to the original source.
198
- - Document negative results: sources searched that did not yield relevant information.
199
-
200
- ## References
201
-
202
- - Booth, A., Sutton, A., & Papaioannou, D. (2016). *Systematic Approaches to a Successful Literature Review* (2nd ed.). Sage.
203
- - Forte, T. (2022). *Building a Second Brain*. Atria Books.
204
- - Machi, L. A. & McEvoy, B. T. (2016). *The Literature Review* (3rd ed.). Corwin Press.