Category
Developer Tools
Developer-focused skills for code editing, version control, CI/CD pipelines, debugging, and development workflows. These tools help AI agents assist with software engineering tasks.
1,750 skills in this category
knowledge-distillation
davila7/claude-code-templates
83Compress large language models using knowledge distillation from teacher to student models. Use when deploying smaller models with retained performance, transferring GPT-4 capabilities to open-source models, or reducing inference costs. Covers temperature scaling, soft targets, reverse KLD, logit distillation, and MiniLLM training strategies.
SKILLPASSING3 Findingsabout 2 months agoawq-quantization
davila7/claude-code-templates
83Activation-aware weight quantization for 4-bit LLM compression with 3x speedup and minimal accuracy loss. Use when deploying large models (7B-70B) on limited GPU memory, when you need faster inference than GPTQ with better accuracy preservation, or for instruction-tuned and multimodal models. MLSys 2024 Best Paper Award winner.
SKILLPASSING3 Findingsabout 2 months agooutlines
davila7/claude-code-templates
83Guarantee valid JSON/XML/code structure during generation, use Pydantic models for type-safe outputs, support local models (Transformers, vLLM), and maximize inference speed with Outlines - dottxt.ai's structured generation library
SKILLPASSING3 Findingsabout 2 months agoneuropixels-analysis
davila7/claude-code-templates
83Neuropixels neural recording analysis. Load SpikeGLX/OpenEphys data, preprocess, motion correction, Kilosort4 spike sorting, quality metrics, Allen/IBL curation, AI-assisted visual analysis, for Neuropixels 1.0/2.0 extracellular electrophysiology. Use when working with neural recordings, spike sorting, extracellular electrophysiology, or when the user mentions Neuropixels, SpikeGLX, Open Ephys, Kilosort, quality metrics, or unit curation.
SKILLPASSING3 Findingsabout 2 months agokegg-database
davila7/claude-code-templates
83Direct REST API access to KEGG (academic use only). Pathway analysis, gene-pathway mapping, metabolic pathways, drug interactions, ID conversion. For Python workflows with multiple databases, prefer bioservices. Use this for direct HTTP/REST work or KEGG-specific control.
SKILLPASSING3 Findingsabout 2 months agoaudiocraft-audio-generation
davila7/claude-code-templates
83PyTorch library for audio generation including text-to-music (MusicGen) and text-to-sound (AudioGen). Use when you need to generate music from text descriptions, create sound effects, or perform melody-conditioned music generation.
SKILLPASSING3 Findingsabout 2 months agodeepchem
davila7/claude-code-templates
83Molecular machine learning toolkit. Property prediction (ADMET, toxicity), GNNs (GCN, MPNN), MoleculeNet benchmarks, pretrained models, featurization, for drug discovery ML.
SKILLPASSING3 Findingsabout 2 months agodrugbank-database
davila7/claude-code-templates
83Access and analyze comprehensive drug information from the DrugBank database including drug properties, interactions, targets, pathways, chemical structures, and pharmacology data. This skill should be used when working with pharmaceutical data, drug discovery research, pharmacology studies, drug-drug interaction analysis, target identification, chemical similarity searches, ADMET predictions, or any task requiring detailed drug and drug target information from DrugBank.
SKILLPASSING3 Findingsabout 2 months agodistributed-llm-pretraining-torchtitan
davila7/claude-code-templates
83Provides PyTorch-native distributed LLM pretraining using torchtitan with 4D parallelism (FSDP2, TP, PP, CP). Use when pretraining Llama 3.1, DeepSeek V3, or custom models at scale from 8 to 512+ GPUs with Float8, torch.compile, and distributed checkpointing.
SKILLPASSING3 Findingsabout 2 months agounsloth
davila7/claude-code-templates
83Expert guidance for fast fine-tuning with Unsloth - 2-5x faster training, 50-80% less memory, LoRA/QLoRA optimization
SKILLPASSING3 Findingsabout 2 months agomoe-training
davila7/claude-code-templates
83Train Mixture of Experts (MoE) models using DeepSpeed or HuggingFace. Use when training large-scale models with limited compute (5× cost reduction vs dense models), implementing sparse architectures like Mixtral 8x7B or DeepSeek-V3, or scaling model capacity without proportional compute increase. Covers MoE architectures, routing mechanisms, load balancing, expert parallelism, and inference optimization.
SKILLPASSING3 Findingsabout 2 months agohypogenic
davila7/claude-code-templates
83Automated hypothesis generation and testing using large language models. Use this skill when generating scientific hypotheses from datasets, combining literature insights with empirical data, testing hypotheses against observational data, or conducting systematic hypothesis exploration for research discovery in domains like deception detection, AI content detection, mental health analysis, or other empirical research tasks.
SKILLPASSING3 Findingsabout 2 months ago