A small marketplace of Claude Code plugins, each shipping one skill.
In Claude Code, add this repo as a marketplace, then install whichever plugin(s) you want:
/plugin marketplace add QuantumBFS/claude-code-skills
/plugin install download-papers@quantum-bfs
/plugin install submit-slurm-job@quantum-bfs
/plugin install verify-references@quantum-bfs
/plugin install paper-review-checklist@quantum-bfs
/plugin install distill-feedback-from-history@quantum-bfs
The @quantum-bfs suffix matches the name field in .claude-plugin/marketplace.json.
To update later:
/plugin marketplace update quantum-bfs
Download an academic paper as a PDF given a URL, DOI, paper title, or arXiv ID. Resolves in order: open web (author homepage, repositories) → Sci-Hub mirrors → arXiv.
Triggers on phrases like "download paper", "get me the PDF", "fetch this DOI". No configuration required.
Generate and submit sbatch scripts for GPU compute jobs on a Slurm cluster. Handles partition selection, explicit GPU model in --gres, full Python path, and log file management.
Required configuration — add to your project's CLAUDE.md:
## Slurm Configuration
- PYTHON_PATH: `/path/to/miniconda3/envs/myenv/bin/python3`
- PROJECT_DIR: `/home/your_user/private/homefile` # scripts/logs must live here
- PARTITION: `home` # compute partitionInvoke with /submit-slurm-job (or just describe the job — Claude will pick up the skill).
Verify a BibTeX bibliography against CrossRef metadata. The skill:
- Scans your
.texfiles to flag uncited entries. - Queries CrossRef for every entry with a DOI and writes a side-by-side
.bib. - Compares the two and produces a Markdown report listing field-level discrepancies (title, author, journal, volume, pages, year), sorted by similarity.
Useful before submission to catch hallucinated or stale BibTeX. Invoke by asking to "verify references" / "check the bib file". Requires Python 3 and an internet connection (for the CrossRef API).
Review a LaTeX research-paper draft against a fixed checklist (notation, figures, citations, structure, equations, style, reproducibility) and insert inline author-comment reminders at the locations that fail each check.
The skill reads the entire .tex, runs programmatic checks (broken \ref/\cite, unreferenced equation labels, symbol-synonym pairs, one-use acronyms), then walks the checklist category by category and inserts short \lw{...} notes (or \todo{...} — whichever author-comment macro the project's preamble defines). Ends with a per-category summary and the top three most impactful issues.
Invoke by asking to "review my paper" / "apply the paper-review checklist to file X".
Mine your own local Claude Code session history (~/.claude/projects/) for recurring behavioral feedback — corrections, format/voice edicts, error catches, domain-knowledge injections — and distill each pattern into the right kind of artifact:
- CLAUDE.md addition (default) — the natural home for discipline rules and preferences.
- Auto-memory feedback entry — alternative for the same content if you prefer the harness memory system.
- SKILL.md draft — only when the pattern describes a parameterizable multi-step workflow. Rare; behavioral patterns are not workflows.
The pipeline runs in eight phases: extract → tag (regex) → domain-injection (LLM) + filter → cluster → classify into a bucket → distill (autonomous or A/B/C/D/E gate) → render → hand back. Default scope is all projects, last 30 days.
Invoke by asking to "distill recurring feedback from my sessions" / "what should I add to CLAUDE.md?" / "find behavioral patterns I should write down". Single-user, local-only; reads JSONLs and writes drafts to ./distilled/, no network calls, no auto-install.
Note: This plugin replaces the earlier extract-skills-from-history, which mis-rendered behavioral rules as standalone SKILL.md files. The new name and bucketed output reflect that most patterns mined this way are memories, not skills.
.
├── .claude-plugin/
│ └── marketplace.json # marketplace manifest, lists all plugins
└── plugins/
├── download-papers/
│ ├── .claude-plugin/plugin.json
│ └── skills/download-papers/SKILL.md
├── submit-slurm-job/
│ ├── .claude-plugin/plugin.json
│ └── skills/submit-slurm-job/SKILL.md
├── verify-references/
│ ├── .claude-plugin/plugin.json
│ └── skills/verify-references/
│ ├── SKILL.md
│ ├── check_unused_refs.py
│ ├── compare_refs.py
│ └── download_crossref.py
├── paper-review-checklist/
│ ├── .claude-plugin/plugin.json
│ └── skills/paper-review-checklist/SKILL.md
└── distill-feedback-from-history/
├── .claude-plugin/plugin.json
└── skills/distill-feedback-from-history/
├── SKILL.md
└── scripts/
├── extract_turns.py
├── tag_turns.py
└── write_artifact.py
Helper scripts shipped alongside a SKILL.md are referenced from the skill instructions via ${CLAUDE_SKILL_DIR}, which Claude Code expands to the skill's install directory at invocation time.
Add a new plugin by:
- Create
plugins/<your-plugin>/.claude-plugin/plugin.json(use one of the existing files as a template). - Add the skill at
plugins/<your-plugin>/skills/<your-plugin>/SKILL.mdwith YAML frontmatter (name,description). - Append an entry to
.claude-plugin/marketplace.json.
Keep skills well-scoped, parameterize any user-specific paths, and document required configuration in this README.
MIT