| 88 | Since the calls are so simple to integrate (anything able to send an http request can use it), it's also easy to compose single LLM actions to more complex operations. We have a demo based on this technology summarizing the above text and then create a valid DITA map out of it. |
| 89 | |
| 90 | == Some things we're still working on |
| 91 | The software is a great help already, but there are still some gaps to fill: |
| 92 | * Chat is not yet implemented. It basically works like single-step operations, but it must maintain the history in a user context and re-inject it to the AI. |
| 93 | * Currently, briefing data must be provided in the model's native format, for example, a certain JSON structure for ChatGPT 3.5. We will provide an automatic conversion from DITA (with certain conventions) into the model's native format. DITA can then be edited with an XML editor or a specific briefing data editor. |
| 94 | * Optional integration with [wiki:Public/StartPageCinnamon Cinnamon] or other CMS / DMS systems would enable running the briefing development from there, including versioning and lifecycles on the briefings. |
| 95 | * Permission and logging system to restrict use of certain functions and keep track of API cost. |
| 96 | |
| 97 | |
| 98 | == Conclusion |
| 99 | The approach has many advantages: |
| 100 | * It hides briefings from users who just want a result. |
| 101 | * It hides software details from training designers. |
| 102 | * It hides AI models API secrets from the users, they use local secrets instead. |
| 103 | * It standardizes briefings, users just pass data. |
| 104 | * It is flexible - writing a briefing is done at a small fraction of the time and cost of training. |
| 105 | * It can use various models, each for its area where it performs best. |
| 106 | |
| 107 | I also believe that this approach becomes more and more important and remove the need for training in most cases: |
| 108 | * Today's leading LLMs have impressive logical and linguistic capabilities (including software code) and are "smart enough" to learn new tasks from a small briefing. |
| 109 | * Upcoming models will have better reasoning and make more out of the same briefing data. |
| 110 | * Upcoming models allow larger token lengths, and thus, more briefing data (for example, GPT 3.5 was limited to 4k tokens, whereas GPT 4 has an option of 32k tokens, or eight times the text length). |
| 111 | * Software solutions like **djAI** allow breaking down tasks into small steps, making them more accessible to the briefing approach. |
| 112 | |
| 113 | What do you think? [wiki:Public/GenContact Please let me know about your thoughts]. |
| 114 | |
| 115 | |