Skip to main content

Applying fork/join model and git-style cherry picking to AI conversations

In this post we introduce using the fork/join concurrency pattern combined with the git-style cherry picking to manage and use LLM conversations.  

These are loose conceptual connections, so please put down the pitchfork (pun intended).

Just as in git you can cherry-pick a commit from one branch to create a new head on another branch, you can pick conversations across sessions (including between different models) to build context to drive LLM conversations.

You can also create a fork of a session, this creates a parallel set of messages that are detached from, ie forked off the main session within which they run.

You can create new messages that join an existing fork.

This is a nice capability, the degree to which it will prove useful is yet to be determined. One of the possibilities is giving the user an ability to create long term "memory" from AI conversations and being able to easily reuse that memory (set of messages) without the need to rebuild context.

The ability to label forks is especially helpful in letting user know what the fork represents and thus offers the potential of greater utility. Below is a demo of this feature:

 


 

Comments

Popular posts from this blog

Declarative Programming With AI/LLMs

  What Is Declarative Programming Broadly speaking, there are two ways to program/instruct a computer to perform a task, they are imperative vs declarative programming. Imperative programming is what we do the most, we write all the code necessary for the computer to perform a task such that the only thing left for the computer to do is fetch and execute CPU instructions. If you are using Java,C#, Javascript...etc you are doing imperative programming. Declarative programming is a higher-order form of programming, we instruct the computer to perform a task but otherwise let it "figure out" how to do it. Declarative programming requires some sort of software based execution engine. Whereas with imperative programming our code gets compiled into some machine/byte code and then run by the CPU, declarative programming requires a layer of software that does the "magic" that allows you to use it without having to write the precise logic for completing tasks.  I would guess...

Integrating AI CodeGen With Low Code Application Development

This is a demo showing Solvent-Lowcodr working with AI Code generation (ChatGPT) to build web apps in a low code fashion. This demo uses Vuetify Component Library. This is the first in a series showing various ways that AI CodeGen and low code can be combined in Solvent-Lowcodr to accelerate web application development. In subsequent posts we'll show how to use AI CodeGen to build low code building blocks in Solvent-Lowcodr. We'll also show demos for React.

Integrating AI CodeGen With Low Code Application Development - Part 2

In this post we follow up from the previous AI CodeGen integration into a Lowcode environment, which was a manual process.  In this post we showcase the new direct AI integration support, then follow that up with using that integration to automate the process of issuing a prompt and generating a Lowcode app. Part 1 - Basic AI Integration Demo (OpenAI, Anthropic, Gemini) Part 2 -CodeGen and Lowcode Integration via automation using all three models.