Interact with GPT AI models as a power user.
Supports multiple AI providers
- OpenAI GPT3.5/4
- Anthropic Claude 1/2
- HuggingFace Inference API
- Google Generative Language API i.e PaLM API
- All models supported by LLaMA.cpp. Current list:
- LLaMA 🦙
- LLaMA 2 🦙🦙
- Alpaca
- GPT4All
- Chinese LLaMA / Alpaca and Chinese LLaMA-2 / Alpaca-2
- Vigogne (French)
- Vicuna
- Koala
- OpenBuddy 🐶 (Multilingual)
- Pygmalion 7B / Metharme 7B
- WizardLM
- Baichuan-7B and its derivations (such as baichuan-7b-sft)
- Aquila-7B / AquilaChat-7B
Rich prompt engineering support
- Inbuilt pre-defined prompts providing large set of features
- Refactor selection
- Generate unit test
- Complete
- Explain code
- Generate Documentation
- Find problems
- Optimize selection
- Openly available prompt files that can be tweaked as needed.
- Prompts can be enriched with custom or predefined functions that can be engineered and fine-tuned to meet specific user needs.
- Prompts can be saved and used directly within VSCode
- Modify request parameters for GPT APIs as needed
- Post-process response via responseHandlers in prompts.
- Inbuilt pre-defined prompts providing large set of features
- Chat activity bar interface for request/response interaction
- Load/save conversations from history
- Export conversations to a file
- Copy/Insert/Create new files out of GPT response.
- Ask FlexiGPT via editor/command context (select and right-click)
- Ask FlexiGPT via command palette control (ctrl/cmd + shift + p)
- Chat activity bar interface for request/response interaction
Search Stack Overflow from within the editor
Invoke pre-cooked custom CLI commands from within your editor