Hexa PA has been updated.

  • v0.3.0 is a major update, that comes with fixes after an API update broke it, available models updated to the latest gpt-3.5-turbo-0125 and gpt-4-turbo-preview, and the context management has been bufeed.

  • When set to GPT 3.5 Turbo, Max Input Tokens = 0 (unlimited), Max Context Messages > 0, Max Output Tokens > 0 it switches seemlessly between 4k(latest) and 16k(legacy) models depending on how long the messages are.

  • It's no longer an even split between input an output tokens, it can now take advantage of GPT-4 Turbo's 128K context even though it only does 4k output. (Successfully tested with more then 4k context.)