Category: Application Programmer Interfaces (APIs)
-
The Problem with Data
Everyone has the same problem, and its name is data. Nearly every business functions in one of two core data models:
- ERP-centric: One large enterprise system (like SAP, NetSuite, or Microsoft Dynamics) acts as the hub for inventory, customers, finance, and operations. It’s monolithic, but everything is in one place.
- Best-of-breed: A constellation of specialized tools – Salesforce or HubSpot for CRM, Zendesk for support, Shopify or WooCommerce for commerce, QuickBooks for finance – all loosely stitched together, if at all.
In reality, most businesses operate somewhere in between. One system becomes the “system of truth,” while others orbit it, each with its own partial view of the business. That setup is manageable until AI enters the picture.
AI is data-hungry. It works best when it can see across your operations. But ERP vendors often make interoperability difficult by design. Their strategy has been to lock you in and make exporting or connecting data expensive or complex.
That’s why more organizations are turning to data lakes or lakehouses, central repositories that aggregate information from across systems and make it queryable. Platforms like Snowflake and Databricks have grown quickly by helping enterprises unify fragmented data into one searchable hub.
When done well, a data lake gives your AI tools visibility across departments: product, inventory, sales, finance, customer support. It’s the foundation for better analytics and better decisions.
But building a good data lake isn’t easy. I joke in my book The AI Evolution, a bad data lake is just a data swamp, a messy, unstructured dump that’s more confusing than helpful. Without a clear data model and strategy for linking information, you’re just hoarding bytes.
Worse, the concept of data lakes was designed pre-AI. They’re great at storing and querying data, but not great at acting on it. If your AI figures out that you’re low on Product X from Supplier Y, your data lake can’t place the order; it can only tell you.
This is where a new approach is gaining traction: API orchestration. Instead of just storing data, you build connective tissue between systems using APIs, letting AI both see and do across tools. Think of it like a universal translator (or Babelfish): systems speak different languages, but orchestration helps them understand each other.
For example, say HubSpot has your customer data and Shopify has your purchase history. By linking them via API, you can match users by email and give AI a unified view. Better yet, if those APIs allow actions, the AI can update records or trigger workflows directly.
Big players like Mulesoft are building enterprise-grade orchestration platforms. But for smaller orgs, tools like Zapier and n8n are becoming popular ways to connect their best-of-breed stacks and make data more actionable.
The bottom line: if your data lives in disconnected systems, you’re not alone. This is the reality for nearly every business we work with. But investing in data cleanup and orchestration now isn’t just prep, it’s the first step needed to truly unlock the power of AI.
That’s exactly why we built the AI Accelerator at PerryLabs. It’s designed for companies stuck in this in-between state where the data is fragmented, the systems don’t talk, and the AI potential feels just out of reach. Through the Accelerator, we help you identify those key data gaps, unify and activate your systems, and build the orchestration layer that sets the stage for real AI performance. Because the future of AI isn’t just about having the data—it’s about making it usable.
-
Jonathan Turkey
Looking to chat with Jonathan Turkey a conversational AI agent? You should see a widget floating to the bottom right of this web page with a button that says “Gobble Gobble”, click that and enjoy!
-
Introducing my AI Playground and Lab
I’m excited to open up my little corner of the web I’ve been tinkering with – an AI sandbox to easily compare and play with various conversational assistants and generative AI models. This web app, located at labs.jasonmperry.com, provides a simple interface wrapping API calls to different systems that keeps experimentation tidy in one place.
Meet the AI Assistants
Last year, OpenAI released AI Assistants you can train as bots accessing files and calling functions through Retrieval-Augmented Generation (RAG). To test capabilities, I created personalities to check how well these features work for customer service or business needs.
Each of these work assistants works at the fictional firm Acme Consulting, and I uploaded to each bot a company primer detailing the history, leadership, services, values, etc., as a reference. The bots include:
- IT manager, Zack “Debugger” Simmons, is here to help with helpdesk inquiries or to suggest best practices and can help troubleshoot issues or explain configurations.
- HR Coordinator Tina “Sunbeam” Phillips is armed with general HR knowledge and a fictional employee handbook with policy details she can cite or reference. Ask her about the holiday schedule and core schedule or for benefits advice.
- Support Coordinator, Samantha “Smiles” Miles is part of the Managed Services team and helps maintain support tickets in the Jira Service Desk for all of our corporate clients. By using RAG, you can ask for updates on support tickets she can grab with phrases like “Tell me what tickets I have open for Microsoft” or “Get me the status of ticket MS-1234” which call mock endpoints.
In addition to the Acme workers, I wanted to experiment with what an assistant powering something like Humane’s upcoming AI pin might function like; after all, we know that the product makes heavy use of OpenAI’s models.
- The witty assistant Mavis “Ace” Jarvis is trained with a helpful instruction set and some RAG operations that allow her to get the weather or check stock prices. She can also show locations on a map based on a query. Try asking her, “Will the weather in Las Vegas be warm enough for me to swim outside?” or “Nvidia is on a tear, how’s the stock doing today?”
Finally, I used Anthropic’s Claude to create backgrounds for three fictional US political commentators with different background stories. You can get political insight, debate, or get views on current issues from Darren, the Conservative, progressive Tyler, and moderate Wesley. In the wake of a push to create AI that bends to different philosophies, I figured these assistants could offer a view into how three distinct personalities might respond to similar prompts while all trained on the same core data.
Text Generation
Compare multiple models’ outputs side-by-side – currently supporting Cohere, Jurassic, Claude, and ChatGPT. Specify max length, temperature, top p sampling, and more for more tailored responses. I plan to continually add the latest models as they become available for testing how phrasing, accuracy, creativity etc. differ when asking the same prompt.
Image Generation
Similarly, visually compare image results from DALL-E and Stable Diffusion by entering identical prompts. The interpretation variance based on the artists and datasets used to train each is intriguing.
Of course, as a playground and lab, I’m continually adding features and experiments, and I plan to add video generation, summarizers, voice cloning, etc. So check back for the latest or suggest additions.
-
Reddit and end of Open API’s
This sucks. Apollo has been my go-to Reddit reader, and I don’t want that to change, but come the end of the month, it’s happening if I like it or not.If you’re not in the loop on Reddit’s API drama, the TL;DR is Reddit moved from open and free APIs to a fee-based system that charges based on the number of API calls you make. If this sounds familiar, it’s because Twitter went down a similar path, and many other open platforms have decided to shut the doors to open API access. The argument for why? OpenAI and AI models are being trained on hordes of open Internet data and, of course, the possibility to eek some revenue out of all the folks hooked on Reddit’s content.
As you might imagine, that approach makes the cost to run something like Apollo unsustainable.
Is Reddit wrong? Apps built for platforms like Twitter and Reddit are like symbiotic bacteria, but one organism is much more dependent on the other. As a platform, Reddit is about user-generated content, and as with Twitter or LinkedIn, it makes us feel like investors or partners in this whole social sharing experiment. But let’s be honest. Revenue and control of the platform you own is what this is really about. If you’re not in control of the last mile, you can’t control how your consumers interact with you or it. You’re constantly limited in how you can advertise, how you personalize, and the ways you can use them to generate revenue.
Hey Reddit, when you fix the mobile and iPad apps, call Mindgrub. We make great mobile apps.
-
Open API’s
The idea of open APIs and access to platforms has become a surprisingly divisive thing. Like most stories in 2023, our story of APIs starts with Elon Musk and Twitter and the decision to shut down third-party app access.
Many, many, many folks were upset that Twitter would shut off access to TweetBot or Twitterific. These apps have been part of Twitter from the start and one inspired Twitter’s logo. To add insult to injury, this made us all collectively realize that Twitter’s mobile app is not great (call me, Mindgrub builds excellent apps). But, Twitter didn’t just band third-party apps – what it did instead is rate limit API calls and implement a new system to charge based on the amount of API calls per month. The price tag was so hefty that bit by bit folks said nope.
Unrelated to Twitter, OpenAI blew through the doors of technology like the Kool-Aid man. Whoever had AI on the 2023 Bingo card deserves all the money. The products like DALI-2 and ChatGPT continue to blow all of our socks off – but then the deep dark secrets of OpenAI and other AI platforms began to drip out.
These LLM (Large Language Model) systems need data, and when I say data, they need all the data. The more you can feed the dang thing, the better – it’s like Seymore, and it wants all the information. Some of this best information came from the most open of sources, places like Twitter, Reddit, and Stack Overflow. These platforms are unique in havings tons of experts who share their advice or answer questions in the most open forums.
Elon Musk and Twitter responded that this was why they needed to lock down APIs and tweets so that eager AI training models won’t try to consume this valuable training data without paying the troll toll. Reddit and other sources of these models followed, and now we find ourselves full circle.
Apollo, my preferred Reddit reader and the only reader with a pixel pet finds itself facing the same issue as Tweetbot and Twitteriffic, the costs for the APIs they need to use from Reddit cost too much.
I get it. I understand it. But sometimes I think of the founding of our great Internet and a time when information was free and people linked to link for love. I guess that was the Internet’s 70’s – and today is a differnet time, but I can’t help but wonder if the hordes of people training models on the open Internet might find the reigns getting pulled a little tighter. I also wonder if this just continues the trend of pay walls popping up everywhere.