Skip to main content

3 posts tagged with "automation"

View all tags

Building the Software 'Gigafactory'

· 2 min read

1. Outcome-Oriented: Autonomous Debugging

Deliver results, not processes. AI must possess a complete closed-loop capability, from vulnerability detection to self-healing. Whether it is invoking curl for diagnostics or parsing logs, the AI should resolve faults independently and generate test cases to verify its own correctness. Managers focus solely on the final output, never intervening in the intermediate logic.

2. Efficiency Metric: Token-Measured Productivity

Consumption equals output. Redefine productivity: the volume of tokens consumed per month is the sole hard metric for efficiency. Achieve exponential leaps in productivity by measuring the number of $200/mo subscriptions a single person can exhaust or the scale of Agent clusters they can simultaneously drive.

3. Drive Mode: Proactive Autonomy

Break the "Command-Response" loop. Top-tier AI systems should not wait for a human wake-up call. They must possess the capacity for autonomous observation, decision-making, and execution, continuously creating value during "vacuum periods" when no human supervision is present.

4. Fault-Tolerant Design: Order within Chaos (Resilient Architecture)

Constrain flexibility through architecture. Construct a "high-fault-tolerance" underlying architecture. Even if the AI "goes rogue" within local logic, it remains confined within the safety zones of a robust systemic framework. Good architecture grants the AI the freedom to fail without letting the entire system collapse.

5. Asset Form: Modular Progress

Capabilities as assets. AI capabilities must be digitized, measurable, and evolvable. Through modular design, ensure every newly developed capability can be reused and combined like building blocks, forming an ever-accumulating competitive moat.

6. Boundary Expansion: Omni-Agent Factory

Limitless substitution. Squeeze every drop of potential out of AI—from code writing and video production to automated social media management. The goal is to transform the company into a highly automated "Gigafactory," where humans serve as the Chief Architects.

7. Evolutionary Logic: Invent and Simplify

Working backwards; breakthrough via brute force. Do not pay the tax of over-engineering. First, invent the product that serves the customer using the most direct (even "clunky") methods. Once the business loop is validated, utilize technical refinement to simplify and reconstruct.

The Promise and Pain of AI Sales Development Representatives: A Field Report

· 5 min read

In the relentless chase to optimize sales pipelines, AI Sales Development Representatives (AI SDRs) have become one of the buzziest tools of 2025. They promise to automate prospecting, personalize outreach at scale, and drop qualified meetings onto your calendar—without the traditional headcount.

But are they actually delivering?

After talking to dozens of sales leaders and digging through hundreds of reviews across G2, Reddit, and Slack communities, I found a more complex story behind the hype.

AI Sales Development Representatives

The 11x Problem: High Expectations, Mixed Results

11x.ai has become the poster child of this category, claiming to make SDRs “11 times more productive.” It’s a bold promise—and one that sets the bar high.

“I expected the AI to research each prospect like a junior rep would,” one sales director told me. “But all I got were Mad Libs with company names filled in.”

This wasn’t an outlier. Across forums and customer chats, a common theme emerged: the emails feel automated, templated, and often too generic to land.

And when leads reply? The AI often stumbles. As one Reddit user put it:

“It can blast emails all day, but the moment someone says something unexpected, it short-circuits.”

This leaves a strange handoff experience—where prospects believe they’re chatting with a human, only to feel the switch when an actual rep steps in mid-convo.

What’s Actually Working

Despite the frustrations, there are places where AI SDRs shine:

  • Outreach volume: Teams consistently report a massive jump in top-of-funnel activity. One European team told me they now “run outreach 24/7” across time zones thanks to their AI reps.
  • Prospecting help: Tools like 11x.ai do a decent job sourcing leads. “The contact lists it finds are better than expected,” said one German user.
  • Personality insights: Humantic AI impressed several teams with surprisingly accurate personality profiles. “It’s like having a cheat code for the first call,” said a G2 reviewer.
  • Real-time coaching: Cresta takes a different approach—coaching human SDRs in real-time rather than replacing them. It’s especially useful for onboarding new reps or improving call quality without hiring a full-time trainer.

Beyond Performance: Hidden Pain Points

Go past the functionality issues, and deeper structural problems start to surface:

  • Locked-in contracts: Most platforms require 35,00035,000–60,000/year commitments with minimal ways to try before buying. “We’re stuck with a tool that doesn’t work for us,” said one buyer.
  • Technical hiccups: From bugs to laggy dashboards, users—especially in Europe—report reliability issues that break workflows.
  • Customization limits: If your audience is niche or your messaging complex, AI often struggles. “We tuned it for weeks,” said a B2B SaaS exec. “The emails still felt generic.”
  • Data security worries: With sensitive customer data flowing through these systems, several larger companies voiced concerns over how their information might be used—or reused.

The Strategic Dilemma: Build, Buy, or Augment?

Given the trade-offs, sales leaders are approaching AI SDRs in one of three ways:

  • The All-In Crowd: Typically fast-moving, high-volume orgs that prioritize scale. They’re willing to accept AI’s rough edges.
  • The Augmenters: Teams using AI to support (not replace) reps. They use tools like Regie.ai for drafting emails, Humantic for call prep, and keep humans in control of conversations.
  • The DIY Builders: Tech-savvy orgs building custom workflows on top of GPTs and internal data. It’s more work, but gives them control and avoids vendor lock-in.

What Needs to Improve

To move from “interesting” to indispensable, AI SDR vendors need to make real progress on a few fronts:

  1. Handle conversations, not just intros – The biggest gap is follow-through. If AI can’t respond naturally, the illusion breaks.
  2. Go beyond templates – True personalization should reference real business context, not just job titles and company names.
  3. Make pricing more flexible – Teams want to experiment before committing six figures.
  4. Fix the UX – Better onboarding, faster load times, and fewer bugs will go a long way.
  5. Allow deeper customization – Give companies tools to teach the AI their value props, messaging frameworks, and product nuance.

Where This Is Headed

The market seems to be splitting into two directions:

  • Vertical AI SDRs: Industry-specific tools trained on healthcare, finance, or manufacturing language, workflows, and regulations.
  • Lightweight assistants: More affordable tools that support reps with writing, prospecting, and call prep—without pretending to replace them.

The companies that lean into augmentation, not automation, may end up building more sustainable businesses.

The Bottom Line

AI SDRs are a classic example of the enterprise AI hype cycle. The pitch—an infinitely scalable digital sales team—is irresistible. But the reality is still catching up.

For most teams, the smart move today is targeted augmentation: Let AI do what it’s good at—prospecting, drafting, supporting—while keeping humans in the loop for objections, relationship-building, and closing.

Because in sales, as in life, the human touch still matters. Maybe now more than ever.

Have you used AI SDRs? What’s been your experience—worth the hype or too soon to tell?

Enterprise Workflow Agents

· 3 min read

Key Themes and Context

Enterprise Workflows

  • Automation levels range from scripted workflows (minimal variation) to agentic workflows (adaptive and dynamic).
  • Enterprise environments, such as those supported by ServiceNow, involve complex, repetitive tasks like IT management, CRM updates, and scheduling.
  • The adoption of LLM-powered agents (e.g., API agents and Web agents) transforms these workflows by leveraging capabilities like multimodal observations and dynamic actions.

LLM Agents for Enterprise Workflows

  • API Agents
    • Utilize structured API calls for efficiency.
    • Pros: Low latency, structured inputs.
    • Cons: Depend on predefined APIs, limited adaptability.
  • Web Agents
    • Simulate human actions on web interfaces.
    • Pros: Greater flexibility; can interact with dynamic UIs.
    • Cons: High latency, error-prone.

WorkArena Framework

  • Benchmarks designed for realistic enterprise workflows.
  • Tasks range from IT inventory management to budget allocation and employee offboarding.
  • Supported by BrowserGym and AgentLab for testing and evaluation in simulated environments.

Technical Frameworks

Agent Architectures

  • TapeAgents Framework

    • Represents agents as resumable modular state machines.
    • Features structured logs (the "tape") for actions, thoughts, and outcomes.
    • Facilitates optimization (e.g., fine-tuning from teacher-to-student agents).
  • WorkArena++

    • Extends WorkArena with more compositional and challenging tasks.
    • Evaluates agents on capabilities like long-term planning and multimodal data integration.

Benchmarks

  • WorkArena: ~20k unique enterprise task instances.
  • WorkArena++: Focused on compositional workflows and data-driven reasoning.
  • Other tools: MiniWoB, WebLINX, VisualWebArena.

Evaluation Metrics

  • GREADTH (Grounded, Responsive, Accurate, Disciplined, Transparent, Helpful):
    • Prioritizes real-world agent performance metrics.
  • Task-Specific Success Rates:
    • For example, form-filling assistants evaluated at 300x lower cost than GPT-4 through fine-tuned students.

Challenges for Agents in Workflows

  • Context Understanding
    • Enterprise tasks require understanding deep hierarchies of information (e.g., dashboards, KBs).
    • Sparse rewards in benchmarks complicate learning.
  • Long-Term Planning
    • Subgoal decomposition and multi-step task execution remain difficult.
  • Safety and Alignment
    • Risks from malicious inputs (e.g., adversarial prompts, hidden text).
  • Cost and Efficiency
    • Shrinking context windows and modular architectures are key to reducing compute costs.

Future Directions

Augmentation Models

  • Centaur Framework:
    • Separates AI from human tasks (e.g., content gathering by AI, final editing by humans).
  • Cyborg Framework:
    • Promotes tight collaboration between AI and humans.

Unified Evaluation

  • Calls for a meta-benchmark to consolidate evaluation protocols across platforms (e.g., WebLINX, WorkArena).

Advancements in Agent Optimization

  • Leveraging RL-inspired techniques for fine-tuning.
  • Modular learning frameworks to improve generalizability.

Opportunities in Knowledge Work

  • Automation of repetitive, low-value tasks (e.g., scheduling, report generation).
  • Integration of multimodal agents into enterprise environments to support decision-making and strategic tasks.
  • Enhanced productivity through human-AI collaboration models.

This synthesis connects the theoretical and practical elements of enterprise workflow agents, showcasing their transformative potential while addressing current limitations.