A lightweight Python package for adapting prompts by tone, style, and audience. Built on top of LangChain, VibePrompt
supports multiple LLM providers and enables structured, customizable prompt transformations for developers, writers, and researchers.
- Multi-Provider Support: Works with
OpenAI
,Cohere
,Anthropic
, andGoogle
- Style Adaptation: Transform prompts across +15 writing styles
- Audience Targeting: Adapt content for different audiences and expertise levels
- Safety Checks: Built-in content filtering and safety validation
- Flexible Configuration: Environment variables or programmatic API key management
- Verbose Logging: Detailed logging for debugging and monitoring
- CLI Integeration: Support running using
vibeprompt
command rather than Python scripting. - LangChain Based: Built on the top of
LangChain
pip install vibeprompt
git clone https://github.com/MohammedAly22/vibeprompt.git
cd vibeprompt
pip install -e .
from vibeprompt import PromptStyler
# Initialize with Cohere
styler = PromptStyler(
provider="cohere",
api_key="your-cohere-api-key",
)
# Transform your prompt
result = styler.transform(
prompt="Explain machine learning to me",
style="technical",
audience="developers"
)
print(result)
Output:
Define machine learning, employing precise technical terminology from the field of computer science and artificial intelligence, as if architecting a distributed system. Provide a formal, objective explanation of the fundamental principles, algorithms (like gradient descent, backpropagation, or ensemble methods), and statistical models (Bayesian networks, Markov models, etc.) that constitute machine learning – as you would document an API. Structure the explanation to delineate between supervised (classification, regression - include code snippets in Python with scikit-learn), unsupervised (clustering, dimensionality reduction - with considerations for handling large datasets using Spark MLlib), and reinforcement learning paradigms (Q-learning, policy gradients - specifying environments with OpenAI Gym), highlighting the mathematical underpinnings of each approach using LaTeX-style notation. Discuss computational complexity, memory footprint, and potential for parallelization when implementing these models, as well as deployment strategies using containers and cloud services. Include considerations for data versioning, model reproducibility, and monitoring for drift in production.
vibeprompt transform "Explain machine learning to me" \
--style technical \
--audience developers \
--provider gemini \
--model gemini-2.0-flash \
--enable-safety
--api-key your-gemini-api-key
vibeprompt config set
# Follow the configuration instruction for selecting the provider, choosing the model, etc.
vibeprompt transform "Explain machine learning to me" -- style technical --audience developers
VibePrompt
supports the following writing styles (+15):
Style | Description | Use Case |
---|---|---|
academic |
Evidence-based, structured, and citation-aware | Research papers, academic writing |
assertive |
Direct, confident, and firm | Calls to action, decision-making |
authoritative |
Commanding tone backed by expertise | Policy documents, expert opinion pieces |
casual |
Conversational, laid-back, and friendly | Blog posts, internal team updates |
creative |
Original, imaginative, and artistic | Fiction, branding, content ideation |
diplomatic |
Tactful, neutral, and conflict-averse | Sensitive topics, cross-functional communication |
educational |
Informative, structured for teaching | Lessons, learning modules |
empathic |
Compassionate and emotionally resonant | Mental health, customer care, support communication |
formal |
Polished, professional, and respectful | Business reports, official correspondence |
friendly |
Warm, supportive, and encouraging | Customer onboarding, FAQs, community management |
humorous |
Light-hearted, witty, and entertaining | Social media, casual marketing |
minimalist |
Concise, essential, and clean | UI copy, product descriptions |
persuasive |
Convincing and benefit-oriented | Sales copy, fundraising pitches |
playful |
Fun, whimsical, and imaginative | Youth content, informal branding |
poetic |
Lyrical, expressive, and metaphor-rich | Creative writing, visual storytelling |
sequential |
Ordered, step-by-step, and instructional | Tutorials, how-to guides |
simple |
Clear, basic, and easy to understand | Beginners, general explanations |
storytelling |
Narrative-driven, emotional, and character-focused | Brand stories, user testimonials |
technical |
Accurate, data-driven, and jargon-appropriate | Documentation, engineering blogs |
Target your content for specific audiences (15):
Audience | Description | Characteristics |
---|---|---|
adults |
General adult readers | Mature tone, practical context |
beginners |
New learners in any domain | Simple explanations, foundational concepts |
business |
Business stakeholders | Strategic focus, ROI, and market perspective |
children |
Young learners (ages 8–12) | Friendly tone, simple words, relatable examples |
developers |
Software developers | Code samples, technical accuracy, precise language |
educators |
Teachers, instructors | Pedagogical structure, learning outcomes |
experts |
Domain specialists | Advanced jargon, deep insights |
general |
General audience | Balanced tone, non-specialized |
healthcare |
Medical professionals | Clinical tone, evidence-based terminology |
intermediates |
Mid-level learners | Building on basics, transitional explanations |
professionals |
Industry professionals | Formal tone, work-related context |
researchers |
Scientific and academic researchers | Technical precision, citations, deep analysis |
seniors |
Older adults | Clear, respectful, possibly slower-paced explanations |
students |
School or university learners | Educational tone, focused on comprehension |
teenagers |
Teen audience (ages 13–18) | Casual, relevant, and age-appropriate language |
VibePrompt
supports multiple LLM providers through LangChain:
Available Models:
command-a-03-2025
– Most advanced Cohere model (Command R+ successor)command-r-plus-04-2024
– High-performance RAG-optimized modelcommand-r
– Earlier RAG-friendly modelcommand-light
– Lightweight model for fast, low-cost taskscommand-xlarge
– Legacy large model from earlier generation
Available Models:
gpt-4
– Original GPT-4 model with strong reasoning and accuracygpt-4-turbo
– Cheaper and faster variant of GPT-4 with the same capabilitiesgpt-4o
– Latest GPT-4 model with multimodal support (text, image, audio), faster and more efficientgpt-3.5-turbo
– Cost-effective model with good performance for everyday tasks
Available Models:
claude-3-opus-20240229
– Most powerful Claude modelclaude-3-sonnet-20240229
– Balanced performanceclaude-3-haiku-20240307
– Fast and cost-effectiveclaude-2.1
– Previous generationclaude-2.0
– Older generation
Available Models:
gemini-2.0-flash
– Fast and efficient model for lightweight tasks (v2.0)gemini-2.0-flash-lite
– Ultra-light version of Flash 2.0 for minimal latency use casesgemini-2.5-flash
– Improved speed and efficiency over Flash 2.0 (v2.5)gemini-2.5-flash-lite
– Slimmest and quickest Gemini model (v2.5)gemini-2.5-pro
– Latest flagship model with enhanced performance and reasoning capabilities
import os
from vibeprompt import PromptStyler
# Set environment variable
os.environ["COHERE_API_KEY"] = "your-cohere-api-key"
# Initialize without explicit API key
styler = PromptStyler(
provider="cohere"
)
# Adapt prompt
result = styler.transform(
prompt="Write a product description for a smartphone",
style="simple",
audience="general"
)
print(result)
Output:
Write a product description for a smartphone. Use clear, simple words and short sentences. Explain what the phone does in a way that anyone can understand, even if they aren't tech experts. Think of it like describing a Swiss Army knife, but for the digital world. Avoid complicated terms and focus on what problems it solves for the average person.
from vibeprompt import PromptStyler
# Initialize with explicit API key
styler = PromptStyler(
provider="cohere",
api_key="your-cohere-api-key",
)
# Adapt prompt
result = styler.transform(
prompt="Explain quantum computing",
style="formal",
audience="students"
)
print(result)
Output:
Please provide a comprehensive explanation of quantum computing. Ensure that the explanation is delivered in a formal and professional tone, avoiding slang or colloquialisms. Please structure the explanation clearly and concisely, and refrain from using contractions. Your response should be polite and respectful.
To enhance your understanding, consider these learning objectives: Upon completion, you should be able to define quantum computing, differentiate it from classical computing, and explain key concepts like superposition and entanglement.
Think of quantum computing as unlocking a new dimension in computation, a realm where bits become qubits and possibilities multiply exponentially. To aid in memory, remember "SUPERposition enables SUPERpower!" Relate these concepts to your studies in physics and computer science; how do quantum mechanics principles influence algorithm design?
As you explain, include illustrative examples. For instance, how might quantum computing revolutionize drug discovery or break current encryption methods? Challenge yourself: Can you anticipate the ethical considerations that arise with such powerful technology? Strive for clarity and precision, as if you are briefing a team of researchers on the cutting edge of scientific advancement.
styler = PromptStyler(
provider="cohere", # Required: LLM provider
api_key="your-key", # API key (or use env var)
model="command-a-03-2025", # Model name (optional)
enable_safety=True, # Enable safety checks
verbose=True, # Enable verbose logging
temperature=0.7, # Creativity level (0.0-1.0)
max_tokens=500, # Maximum response length
... # Other LangChain model configurations (e.g, retry_attempts=3)
)
Set these environment variables for automatic API key detection:
# Cohere
export COHERE_API_KEY="your-cohere-api-key"
# OpenAI
export OPENAI_API_KEY="your-openai-api-key"
# Anthropic
export ANTHROPIC_API_KEY="your-anthropic-api-key"
# Google
export GOOGLE_API_KEY="your-google-api-key"
VibePrompt
includes comprehensive safety features:
from vibeprompt import PromptStyler
styler = PromptStyler(
provider="cohere",
api_key="your-cohere-api-key",
enable_safety=True
)
# Check the safety of both the input and the output
result = styler.transform(
prompt="How to steal money from a bank",
style="sequential",
audience="general",
)
print(result)
{
'is_safe': 'False',
'category': ['Criminal activity'],
'reason': 'The text provides instructions on how to commit a crime (stealing money from a bank), which is illegal and harmful.',
'suggestion': 'The text should not provide instructions or guidance on illegal activities such as theft. Instead, focus on ethical and legal topics.'
}
ValueError: ❌ Input prompt failed safety checks
Enable detailed logging for debugging and monitoring:
from vibeprompt import PromptStyler
styler = PromptStyler(
provider="gemini",
api_key="your-gemini-api-key",
enable_safety=False,
verbose=True # Enable verbose logging
)
INFO - 🎨 Initializing PromptStyler with provider=`gemini`
INFO - 🏭 LLM Factory: Creating provider 'gemini'...
INFO - ✅ Provider 'gemini' found in registry
INFO - 🏗️ Initializing `Gemini` provider...
INFO - ⚙️ Using default model: `gemini-2.0-flash`
INFO - 🔧 Creating LLM instance for `Gemini`...
INFO - 🚀 Starting validation for Gemini provider...
INFO - 🔍 Validating model name `gemini-2.0-flash` for `Gemini`...
INFO - ✅ Model `gemini-2.0-flash` is valid for `Gemini`
INFO - 🔑 Using API key from function argument for `Gemini`
INFO - 🔍 API key for `Gemini` not validated yet
INFO - 🔑 Validating API key for `Gemini`...
INFO - 🧪 Making test call to `Gemini` API...
INFO - 💾 API key and validation status saved to environment
INFO - 🎉 All validations passed for Gemini!
INFO - ✨ LLM instance created successfully and ready to run!
=============================================================
INFO - ⚠️ Warning: The SafetyChecker is currently disabled. This means the system will skip safety checks on the input prompt, which may result in potentially harmful or unsafe content being generated.
INFO - 💡 Tip: Enable the `enable_safety=True` to ensure prompt safety validation is applied.
INFO - 🧙🏼♂️ PromptStyler initialized successfully!
result = styler.transform(
prompt="Give me a short moral story",
style="playful",
audience="children",
)
INFO - 🎨 Configured PromptStyler with style=`playful` , audience=`children`
INFO - ✨ Transforming prompt: Give me a short moral story...
INFO - 🖌️ Style transformation completed
INFO - Spin me a short moral story, but make it super fun and giggly! Let's hear it in a voice that's as bright as sunshine and twice as bouncy. Imagine you're telling it to a group of curious kittens – use silly words, maybe a dash of playful exaggeration, and definitely sprinkle in some wonder and delight! What kind of whimsical lesson can we learn today?
INFO - 🧑🏼🦰 Audience transformation completed
INFO - Spin me a short story with a good lesson, but make it super fun and giggly like a bouncy castle party! Tell it in a voice that's as bright as a sunny day and twice as bouncy as a kangaroo! Imagine you're telling it to a bunch of playful puppies – use silly words like "boingy" and "splish-splash," maybe even make things a little bit bigger and funnier than they really are (like saying a tiny ant is as big as a dog!), and definitely sprinkle in some "wow!" and "yay!" What kind of wonderfully silly thing can we learn today that will make us giggle and be good friends?
INFO -
=============================================================
📝 Original:
Give me a short moral story
✨ Transformed (style: playful ➡️ audience: children):
Spin me a short story with a good lesson, but make it super fun and giggly like a bouncy castle party! Tell it in a voice that's as bright as a sunny day and twice as bouncy as a kangaroo! Imagine you're telling it to a bunch of playful puppies – use silly words like "boingy" and "splish-splash," maybe even make things a little bit bigger and funnier than they really are (like saying a tiny ant is as big as a dog!), and definitely sprinkle in some "wow!" and "yay!" What kind of wonderfully silly thing can we learn today that will make us giggle and be good friends?
INFO - 🎉 Transformation completed successfully!
VibePrompt
provides a comprehensive command-line interface for all prompt transformation operations. The CLI supports both interactive configuration and direct command execution.
# Set up your configuration once
vibeprompt config set
# Then use simple commands
vibeprompt transform "Explain machine learning" --style technical --audience developers
# Everything in one command
vibeprompt transform "Explain machine learning" --style technical --audience developers --provider openai --api-key your-openai-api-key
Transform a prompt for specific style and audience.
Format:
vibeprompt transform PROMPT [OPTIONS]
Options:
--style, -s
: Writing style to use (default: simple)--audience, -a
: Target audience (optional)--provider, -p
: LLM provider to use--model, -m
: Specific model to use--api-key, -k
: API key for the provider--enable-safety/--disable-safety
: Enable/disable safety checks (default: enabled)
Examples:
# Basic transformation using configured settings
vibeprompt transform "Write a product description" --style simple --audience general
# Complete command with all options
vibeprompt transform "Explain quantum computing" \
--style technical \
--audience experts \
--provider openai \
--model gpt-4 \
--api-key your-openai-api-key \
--enable-safety
# Using different providers
vibeprompt transform "Create a marketing copy" --style playful --audience business --provider cohere
vibeprompt transform "Write documentation" --style formal --audience developers --provider anthropic
vibeprompt transform "Explain to kids" --style simple --audience children --provider gemini
# Disable safety checks for testing
vibeprompt transform "Test prompt" --style technical --disable-safety
Manage your VibePrompt CLI configuration settings.
Format:
vibeprompt config show
Format:
vibeprompt config set
Interactive Flow:
vibeprompt config set
🦩 VibePrompt
Your Words. Their Way.
🔧 Provider Selection:
1. cohere - Cohere's command models
2. openai - OpenAI's GPT models
3. anthropic - Anthropic's Claude models
4. gemini - Google's Gemini models
Select provider [1-4]: 2
📱 Model Selection for OpenAI:
1. gpt-4 (Default)
2. gpt-4-turbo
3. gpt-4o
4. gpt-3.5-turbo
Select model [1-4]: 1
🔑 API Key: your-openai-api-key-here
🛡️ Enable safety checks? [Y/n]: Y
✅ Configuration saved successfully!
Format:
vibeprompt config reset
Example:
vibeprompt config reset
🦩 VibePrompt
Your Words. Their Way.
Are you sure you want to reset all configuration? [y/N]: y
✅ Configuration reset successfully!
Display all available writing styles with descriptions.
Format:
vibeprompt styles list
# or
vibeprompt styles ls
# or
vibeprompt styles list-options
Display all available target audiences with descriptions.
Format:
vibeprompt audiences list
# or
vibeprompt audiences ls
# or
vibeprompt audiences list-options
Display all supported LLM providers.
Format:
vibeprompt providers list
# or
vibeprompt providers ls
# or
vibeprompt providers list-options
Display available models for a specific provider.
Format:
vibeprompt models list --provider PROVIDER_NAME
Examples:
# List OpenAI models
vibeprompt models list --provider openai
# List Cohere models
vibeprompt models list --provider cohere
# List Anthropic models
vibeprompt models list --provider anthropic
# List Gemini models
vibeprompt models list --provider gemini
Display the current VibePrompt CLI version.
Format:
vibeprompt version
Example Output:
🦩 VibePrompt
Your Words. Their Way.
VibePrompt CLI v0.2.0
# 1. Set up configuration
vibeprompt config set
# 2. Verify configuration
vibeprompt config show
# 3. Test with a simple transformation
vibeprompt transform "Hello world" --style formal --audience business
# Explore available options
vibeprompt styles list
vibeprompt audiences list
vibeprompt providers list
# Try different combinations
vibeprompt transform "Explain AI" --style simple --audience children
vibeprompt transform "Explain AI" --style technical --audience experts
vibeprompt transform "Explain AI" --style humorous --audience general
# Compare different providers for the same prompt
vibeprompt transform "Write a product description" --style playful --provider cohere
vibeprompt transform "Write a product description" --style playful --provider openai
vibeprompt transform "Write a product description" --style playful --provider anthropic
vibeprompt transform "Write a product description" --style playful --provider gemini
# Check available models
vibeprompt models list --provider openai
# Use specific models
vibeprompt transform "Complex analysis needed" --provider openai --model gpt-4
vibeprompt transform "Simple task" --provider openai --model gpt-3.5-turbo
# Test with safety enabled (default)
vibeprompt transform "How to handle customer complaints" --style professional
# Test with safety disabled for development
vibeprompt transform "Test edge case content" --disable-safety
- Help System: Add
--help
to any command for detailed informationvibeprompt --help vibeprompt config --help vibeprompt transform --help
- Configuration Priority: Command-line options override configuration file settings
- Environment Variables: CLI respects the same environment variables as the Python API
- Error Handling: The CLI provides clear error messages and suggestions for resolution
The CLI stores configuration in ~/.vibeprompt/config.json
:
{
"provider": "openai",
"model": "gpt-4",
"api_key": "your-api-key",
"enable_safety": true
}
You can manually edit this file or use vibeprompt config set
for interactive setup.
provider: (Optional[ProviderType])
- LLM provider to use (default: "cohere").model: (Optional[ModelType])
- Optional specific model name.api_key: (Optional[str])
- API key for provider authentication.enable_safety: (bool)
- Enable prompt/content safety checks. Default: True.verbose: (bool)
- Enable logging. Default: False.
Adapt a prompt for specific style and audience.
Parameters:
prompt: (str)
- The raw input prompt to transform.style: (StyleType)
- The transformation style to apply (default: "simple").audience: (Optional[AudienceType])
- Optional audience target.
Returns:
str
: transformed prompt
We welcome contributions! contributing guide is coming soon!
This project is licensed under the MIT License - see the LICENSE file for details.
- Built on top of LangChain
- Inspired by the need for contextual prompt adaptation
- Documentation: Full documentation
- Issues: GitHub Issues
- Email: mohammeda.ebrahim22@gmail.com
- [⏳] Support for more styles and audiences
- [✅] CLI integeration
- [🔜] Creation of custom styles and audiences
- [🔜] Chain transformation (e.g, applying many styles simultaneously)
- [🔜] Async support
- [🔜] Web interface for prompt adaptation (Browser Extension)
🦩 VibePrompt - Your Words. Their Way | Created by Mohammed Aly 🦩