PacketLLM 0.1.1
- Models: restrict built-in presets to the OpenAI GPT-5 family
(gpt-5,gpt-5-mini,gpt-5-nano);
remove previous presets (gpt-4o,gpt-4o-mini,gpt-4.1,o1,o3-mini).
- API compatibility: avoid 400 “unsupported_value” by omitting
temperaturefor GPT-5 models (treated as simplified in this
release).
- UI/Settings: remove the temperature control and related logic; all
conversations use the model’s default temperature. Documentation and
tests updated accordingly.
PacketLLM 0.1.0
- Initial Release to CRAN.
- Provides an RStudio Gadget (run_llm_chat_app()) for
interactive chat with OpenAI models (e.g., GPT-4o, GPT-4.1).
- Supports multiple, tabbed conversations.
- Allows uploading .R,.pdf, and.docxfiles as context for models supporting
attachments.
- Enables setting the model, temperature, and system message per
conversation (before the first message).
- Features asynchronous API calls to keep the R console
responsive.
- Includes helper functions for managing conversation history and
settings (create_new_conversation,get_active_chat_history,set_conversation_model, etc.).
- Requires an OPENAI_API_KEYenvironment variable for API
access.
- Added basic unit tests using testthat.
- Added README.mdand an introductory vignette
(vignettes/PacketLLM-introduction.Rmd).