Jul 30, 2024
For the past 70 years, the world of software was one of comprehending and analyzing numbers.
Early ML native applications required extremely sophisticated teams and mountains of clean, highly structured data. They could consume tables, join data, and perform complex numerical analyses. However, akin to a train running on a single track, it was difficult to deviate from a narrowly intended range of motion. Systems were brittle, needing precise connection through APIs. Even quantitative data was of limited utility if not highly structured. Applications could not digest unstructured data such as text, images, audio and video – or express complex ideas that require their usage.
Enter Large Language Models. With the rise of GPUs, cloud computing, and innovation in model architectures, LLMs can ingest and analyze multimodal data (i.e., pictures, videos, voice) to produce integrated responses. They do this at scale while accessing a corpus of knowledge across the web or relevant information within a company’s firewall.
In this piece, we outline some characteristics of AI applications:
Applications can live in the space between structured and unstructured data for the first time.
Multimodal data (e.g., pictures, videos, voice) can build on and amplify the value of structured data. Integrating such data with traditional structured data enables software companies and leveraging the power of LLMs to “reason” on the expanded corpus enables software to radically expand scope and produce outputs at near submission-ready state (though we still strongly recommend a human review cycle.)
Multi-modal ingest can massively improve specialized industry applications
First wave LLM applications emerged with key word aggregation, synthesis and output capabilities. One example is Hebbia’s tool, Matrix, that automates much of a financial analysts’ rote work by surfacing key investment insights from SEC filings, public analyst reports and other written documents.
As models get better, the next tide of applications will connect relationships across diverse data types such as voice, text, and images. Healthcare diagnosis, management, and research is a compelling case where unstructured data (clinical notes, medical imaging) can account for up to 80% of a patient’s medical record (National Library of Medicine).
Several non-clinical AI applications are tackling the enormous administrative burdens of healthcare, such as prior authorization (Anterior, Develop, Latent), coding, billing, and utilization. Akasa accelerates prior auth mentioned above, as well as automates claim attachment and submission to payer portals. Zentist’s Revenue Cycle Management AI solution can fast track the 8+ hours a week Dental Service Organizations spend on manually managing claims.
LLMs can also aid in clinical decisioning support by surfacing medical research to develop diagnoses, while still giving the clinician authority over to accept, edit, or reject the LLM output (Glass Health, Carenostics). AI can analyze large volumes of unstructured molecular, clinical, and genomic data to provide us a more granular level of understanding of the bodies’ response to therapy (Tempus Health). Healthcare companies with ambitions to train proprietary models that use new types of data seek previously untapped health system datasets. Startups are tackling next generation healthcare data businesses, marketplaces (Segmed, Protege) and platforms that speed-up backend processes like cleaning, harmonization and error audit.
Humans no longer need to manually create comprehensive deliverables bit by bit, they review AI outputs
Dashboards, drafts, and other outputs historically required many cycles of human input. Lawyers metabolized hundreds of long documents; plan reviewers surfaced thousands of codes across changing codebooks. Now, AI applications can automate a large percentage of that workflow. Humans add value by shaping, validating, and updating AI outputs.
Companies can reduce time and dependence on manual processing, saving cognitive load and freeing workers to focus on higher value-added processes. EvenUp software learns from a dataset of 250K+ public verdicts and private settlements to auto generate entire demand packages for industry lawyers; Vic.ai uses optical character recognition and natural language processing to cut through hours of accounting processes – even submitting invoices for approval. With TrunkTools search, construction workers ask questions about specifics within project docs such as contract details, what type of home fixtures are present within drawings, and get comprehensive answers back, saving time spent continuously searching and rework.
Human-in-the-loop is still required for most solutions today, but as LLMs improve, feedback on quality and desirability of model outputs via in-context learning will guide the learning process. Human interaction creates a data flywheel that trains the model and deepens a moat of proprietary data advantages.
Early-stage learnings on building with LLMs
Here are some key implications for builders, based on what we’ve experienced in successful AI startups to date:
We’ve already made a dozen investments in and around AI applications. We’re on the lookout for more. If that’s you, let’s chat!