"As well as noting the complexity of building generative AI applications"
Understatement of the year, so far.
For a while now we have been using AI as a replacement app service. It is much better and much worse. The biggest plus is that you define the application. The biggest minus is the LLM corrupting the data. An example is something like this:
Me: New app to track my dog's growth
AI: Sure, give me some details. What breed is he?
Me: Akita/Malinowa cross
AI: Blond Akita
Me: NO, u stupid cow. A--KI-TA cross with MA-LIN-WA
AI: Leopard-pattern coat, Akita Belgium cross
No: Sweet Jesus, women! Where did you get f-ing Leopard-pattern from
AI: I'm sorry blah blah blah for 1 minute
... 30 seconds later and repeated simplified prompts mixed with the kind of language my mother would not think possible of me!
AI: Now I understand.
Me: Stupid B!tch
I wouldn't worry too much about the impact of humans as they already know that most of them are on borrowed time and glad that they have blagged it this far.
We did a model for a law firm in Lincolns' Inns. First meeting I said sack half the para-legal team. She said no chance. Fast forward >> 1 desktop Mint box with a sweet graphics card sitting in the corner running LLM localhost, which does 50% of the work the paralegals did. Hence farewell to about 30% of the team and the huge savings of their wages.
With our current snapshot models, the data integrity problem I mocked above is a no, just no. I've seen all I want to see and it's still no. It is the biggest stumbling block to one app to rule - well no apps in the traditional sense. We need to fix this problem but, how? The core model is so far away from clean data that it is near unusable as we have found in recent months.
The best solution is to brew your own LLM as soon as you can. It is your early website circa 1996. Get it done. LLM models will replace YouTube. Go Android for the best models. F Apple and their steam-powered junk. Run a local LLM on your phone and block internet access for it. You can shift the weights and biases to a new model at anytime.