Get Historical Employee Growth With Proxycurl's Employee Count Endpoint Learn more

sales prospecting with AI agents
"Yeah LLMs are cool and all, but how can they really benefit my workflow?" - you, probably. AI agents are how. Allow me to explain.

/ proxycurl

How to Automate Sales Prospecting & Outreach with AI Agents

Ever since LLMs like ChatGPT have hit the scene, they've been all the rage and have fundamentally changed how many things are done—sales prospecting and outreach included.

But the ultimate question is: how can you practically implement LLMs to improve your workflow, rather than just being something fun to play with?

You know, at what point does "AI" become something like we originally imagined it being—something that can complete entire tasks for you?

For many things, AI isn't quite there yet. But for sales prospecting and outreach, quite a few things can be automated, or at least heavily assisted by AI agents right now.

That's what this article will be all about.

First things first:

What is an AI agent?

Good question. It'll have a slightly different answer depending on who you ask.

For example, IBM describes an AI agent as:

a system or program that autonomously performs tasks for a user or another system by designing its workflow and utilizing available tools.

On the other hand, Amazon defines an AI agent as:

software that interacts with its environment, collects data, and uses it to perform self-determined tasks to meet goals set by humans.

Then TechCrunch describes an AI agent as:

AI-driven software that performs tasks traditionally done by humans, potentially crossing multiple systems and handling various jobs beyond just answering questions.

Basically, at its core, an AI agent is a software program designed to perform tasks autonomously. These agents can process information, make decisions, and execute actions based on predefined rules and learning from data.

How can AI agents benefit sales prospecting?

For sales prospecting, an AI agent can accomplish three main things:

  • Lead identification and ranking: An AI agent can search for leads that fit your Ideal Customer Profile (ICP), identify and rank them, and prospect on your behalf.
  • Automated outreach: The agent can automatically reach out to ideal prospects and write a series of outreach messages tailored to their interests and needs.
  • Inbox management: The AI agent can monitor your email inbox, automatically replying to interested leads and pushing them to book a call with you directly.

For B2B sales, this means you can use AI agents to handle the majority of the sales workflow. All you need to do is show up for sales demos and close the deal.

The first step to any AI agent is data

To get an LLM to do anything useful, it needs an appropriate amount of data for the task at hand.

For example, if you're automating sales prospecting, the AI agent would need access to relevant prospecting data, such as email addresses, company information, and so on.

How to programmatically access B2B data

Luckily since you're on the Proxycurl blog, you'll have programmatic access to just about all of the B2B data you could ever possibly need.

We have millions of data points on people, jobs, companies, and more, and we provide acccess to it all via a REST API.

In this case, two very useful endpoints would be our Person Search Endpoint, and our Personal Email Lookup Endpoint.

Searching for prospects

So, for example, to search for prospects via our Person Search Endpoint, you could do so via a simple cURL command:

curl \
    -G \
    -H "Authorization: Bearer ${YOUR_API_KEY}" \
    'https://nubela.co/proxycurl/api/v2/search/person' \
    --data-urlencode 'country=US' \
    --data-urlencode 'current_role_title=founder' \
    --data-urlencode 'industries=Computer Software' \
    --data-urlencode 'page_size=10' \
    --data-urlencode 'enrich_profiles=enrich'

That would return you 10 "enriched" (more information) founders working in the computer software industry.

💡
Note: You can create your Proxycurl account for free right here and test out a few API queries using the trial credits you'll receive.
You'll get 100 trial credits if you sign up with a work email, 10 credits with a personal email.

Then, if you found a prospect of interest that was returned by our Person Search Endpoint and were interested in reaching out, you could use our Personal Email Lookup Endpoint to get their email.

Enriching prospects with emails

Here's how:

curl \
    -G \
    -H "Authorization: Bearer ${YOUR_API_KEY}" \
    'https://nubela.co/proxycurl/api/contact-api/personal-email' \
    --data-urlencode 'linkedin_profile_url=https://www.linkedin.com/in/exampleprofile/'

That would take the given LinkedIn URL from the Person Search Endpoint, and return you their personal email if available.

Generating large lead lists to act as a dataset for your AI agent

Now let's say you want to generate a large lead list to use as a dataset for your AI agent.

We could use a little bit of Python to help us accomplish this:

import requests
import os
import csv
import time

CSV_FILE_PATH = "./exported_data.csv"
API_KEY = "Your_API_KEY_Here"


# Function to search for people and handle pagination
def search_person(api_key, max_results):
    url = "https://nubela.co/proxycurl/api/v2/search/person"
    headers = {"Authorization": f"Bearer {api_key}"}

    params = {
        'country': 'US',
        'current_role_title': 'founder',
        'industries': 'Computer Software',
        'page_size': '10',
        'enrich_profiles': 'enrich'
    }

    results = []
    try:
        while len(results) < max_results:
            response = requests.get(url, headers=headers, params=params)

            if response.status_code == 429:
                print("Rate limit hit, sleeping for a while...")
                time.sleep(10)
                continue

            response.raise_for_status()
            data = response.json()

            if not data.get("results"):
                print("No more results to fetch.")
                break

            results.extend(data.get("results", []))
            print(f"Fetched {len(results)} results so far.")

            if data.get("next_page") and len(results) < max_results:
                url = data["next_page"]
            else:
                break

        return results[:max_results]
    except requests.exceptions.RequestException as e:
        print(f"Error in search_person: {e}")
        return []


# Lookup personal email function
def lookup_personal_email(api_key, linkedin_url):
    url = 'https://nubela.co/proxycurl/api/contact-api/personal-email'
    headers = {'Authorization': f'Bearer {api_key}'}
    params = {'linkedin_profile_url': linkedin_url}

    try:
        response = requests.get(url, headers=headers, params=params)
        response.raise_for_status()
        data = response.json()
        return ', '.join(data.get('emails', [])) or 'N/A'
    except requests.exceptions.RequestException as e:
        print(f"Error fetching personal email for {linkedin_url}: {e}")
        return 'N/A'


# Create new CSV if needed
def create_new_csv(file_name):
    with open(file_name, mode='w', newline='', encoding='utf-8') as file:
        writer = csv.writer(file)
        writer.writerow([
            "First Name", "Last Name", "LinkedIn URL", "Occupation", "Summary",
            "Current Company", "Current Company Description", "Company URL",
            "Experiences", "Personal Email"
        ])
    print(f"New CSV file created: {file_name}")


# Export data to CSV with handling for missing profile data
def export_to_csv(data, file_name=CSV_FILE_PATH):
    print(f"Starting export to CSV: {file_name}")

    if not os.path.exists(file_name):
        create_new_csv(file_name)

    try:
        with open(file_name, mode='a', newline='', encoding='utf-8') as file:
            writer = csv.writer(file)

            for person in data:
                linkedin_url = person.get("linkedin_profile_url", "N/A")
                profile = person.get("profile", {})

                if not profile:
                    print(f"Profile missing for {linkedin_url}, skipping...")
                    continue

                first_name = profile.get("first_name", "N/A")
                last_name = profile.get("last_name", "N/A")
                occupation = profile.get("occupation", "N/A")
                summary = profile.get("summary", "N/A")

                experiences = profile.get("experiences", [])
                experiences_list = []
                current_company = "N/A"
                current_company_description = "N/A"
                company_url = "N/A"

                if experiences and isinstance(experiences, list):
                    first_experience = experiences[0] if len(experiences) > 0 else {}
                    current_company = first_experience.get("company", "N/A")
                    current_company_description = first_experience.get("description", "N/A")
                    company_url = first_experience.get("company_linkedin_profile_url", "N/A")

                    for exp in experiences:
                        company_name = exp.get("company", "N/A")
                        title = exp.get("title", "N/A")
                        description = exp.get("description", "N/A")
                        start_date = exp.get("starts_at", {}).get("year", "N/A") if exp.get("starts_at") else "N/A"
                        end_date = exp.get("ends_at", {}).get("year", "N/A") if exp.get("ends_at") else "N/A"
                        exp_string = f"Company: {company_name}, Title: {title}, Description: {description}, Start: {start_date}, End: {end_date}"
                        experiences_list.append(exp_string)
                    experiences_str = " | ".join(experiences_list)
                else:
                    experiences_str = "N/A"

                personal_email = lookup_personal_email(API_KEY, linkedin_url)

                writer.writerow([
                    first_name, last_name, linkedin_url, occupation, summary,
                    current_company, current_company_description, company_url,
                    experiences_str, personal_email
                ])
                print(f"Added row for {first_name} {last_name}")

        print(f"Data successfully exported to {file_name}")
    except Exception as e:
        print(f"Error in export_to_csv: {e}")


# Start the process of fetching and exporting data
def start_process(max_results):
    try:
        search_results = search_person(API_KEY, max_results)
        if not search_results:
            print("No results found.")
            return
        export_to_csv(search_results)
    except Exception as e:
        print(f"Error in start_process: {e}")


if __name__ == '__main__':
    start_process(max_results=500)

How to use the script

1. Make sure you have Python installed.

To run the script you would first need to have Python installed. An easy way to work with Python is to install an Integrated Development Environment (IDE) like PyCharm.

Also, install the requests module if you haven’t already using Python's package manager.

Just run the following command from your terminal or command prompt:

pip3 install requests

2. Replace Your_API_Key_Here with your actual Proxycurl API key in the script.

3. Execute the script:

python3 your_script_name.py

4. After running the above Python script, it would return you a .CSV list of 500 founders of computer software companies.

An example of the CSV returned
An example of the CSV returned

The data points it includes are:

  • First name
  • Last name
  • LinkedIn profile URL
  • Occupation
  • Summary (a brief overview of the individual's profile)
  • Current company name
  • Current company description
  • Current company LinkedIn URL
  • Experiences (a summary of job experiences, including company names, job titles, descriptions, start and end dates)
  • Personal email (if available)

For all prospects, and can return even more if you want to add more of the results available on our Person Search Endpoint, which you can view on our documentation here.

Note that this can consume quite a bit of credits, so you might want to review our pricing here first and lower the amount of results returned by the Python script by changing max_results.

Modifying the filtering for the script

You can adjust the search parameters in the params dictionary inside the search_person function.

For example, if you want to search for marketing managers in the UK instead of founders in the US, you can update the parameters like this:

params = {
    'country': 'GB',
    'current_role_title': 'marketing manager',
    'industries': 'Marketing and Advertising',
    'page_size': '10',
    'enrich_profiles': 'enrich'
}

But that's far from it when it comes to filtering and search options. You can view the available search parameters here.

Feel free to modify the parameters to match your Ideal Customer Profile (ICP) based on job titles, industries, locations, or other available filters.

💡
Note: The above script only uses "personal emails," but you'll find that sometimes personal emails as well as work emails will be returned. You can also add work emails to this via our Work Email Lookup Endpoint, but you'll have to configure a webhook as this endpoint functions asynchronously. You could use something like Flask and ngrok to accomplish that locally. I've skipped that for the sake of keeping this article to the point, but it isn't too difficult.

Creating AI agents with Zapier Central

Cool. Now we have a dataset to work with.

For the next step, in order to bring this all together, we'll be using Zapier Central.

Zapier's AI assistants
Zapier's AI assistants

Zapier is well known for integrating systems/apps. It's what allows you to piece together all of the different moving parts of your workflow.

And now they've dipped their toes into AI agents too. Best of all they'll allow you 400 activities per month and live data sources plus web browsing for free.

💡
Note: For a self-hosted option, you could also use n8n along with LangChain & LLAMA 3 to create similar AI automation systems. But there are quite a bit more complexities, and I think you'll find you often don't save that much self-hosting things like this.

After creating your Zapier account, you can access Zapier Central right here.

You'll then see a dashboard similar to below:

Zapier Central's dashboard
Zapier Central's dashboard

Click the "Plus" icon to create a new AI agent:

Creating a new AI agent
Creating a new AI agent

AI agents for lead identification and ranking

Next up you'll need a name and description, which is whatever you'd like. Then the instructions which will apply to every conversation.

In this case, we can use something like, "your job is to be a member of my sales team and help me prospect plus close sales."

Configuring your AI agent
Configuring your AI agent

As of right now, Zapier Central's best way to natively integrate with business suite tools is by using Google's suite of tools. Such as Google Docs and Gmail.

To be honest with you, though, for outreach, Gmail is one of the best options you can use as there are a significant amount of Gmail users, and Gmail would much prefer mail came from their own infrastructure. Google's line of products aren't too expensive.

Anyway, you'll want to authenticate Zapier with your Google account and you'll be able to directly sync.

Then for your actions:

Zapier's AI actions
Zapier's AI actions

Select Gmail "Create draft," and Google Docs "Create document from text."

💡
Note: You can also select "send email" instead of create draft, but this will indeed automatically and directly send the email. It's up to you if you want to review things before they get sent out or not.

Then upload the .CSV full of prospects we generated earlier with Proxycurl as the data source:

Adding your data source for your AI agent
Adding your data source for your AI agent

It should look similar to this:

Everything configured
Everything configured

Then click "Save" and "Create assistant."

Next, open your assitant, and you'll see a page such as this:

Your new AI assistant
Your new AI assistant

Click on "Behaviors" in the middle and then "Create behavior":

Creating your AI agents behaviors
Creating your AI agents behaviors

For the trigger, there are a couple of different options you can choose:

Triggering your AI agent
Triggering your AI agent

Such as Slack, a scheduled time, a Gmail action, or more. In this case I'll simply select when I message the assistant the word "Start."

Next up you'll need to fill out the instructions to follow (prompt) section.

So, for mine, here's what I used:

I work for Proxycurl, a B2B data provider & API. We want to target technical founders that could use our B2B data in their business. I want you to use the .CSV and the data provided to first rank the top 100 prospects and give a brief explanation why, then save them in a Google Document. 

After that, get rid of the ones unlikely to convert, and  draft an outreach email to the rest pitching a simple and short email to book a call. It should utilize the data provided, such as the summary and position/role to personalize the email and convey value.

For each row in the Leads spreadsheet, do ALL of the following steps:
1. Use the name and email to populate contact information. 
2. Use any information available about the contact to create a draft an email. 
3. Generate the body and subject.
4. Sign the email with: "Looking forward to connecting!" and "[your name here]"

Here's an example for the company AgentHub with the contact AgentHub team: 

Hey AgentHub team,

Just checked out your platform, and I'm impressed with how easily one can drag and drop to create AI-driven workflow automations. I'd love to connect and explore  explore how our product could help you unlock more potential.

Here's a link to my calendar if you want to chat this week!
[https://calendly.com/proxycurl]

Of course you'll want to slightly alter it your liking and what fits you and your business.

Then click on the "Test behavior" button:

Seeing how your AI agent will behave based on your instructions
Seeing how your AI agent will behave based on your instructions

It will display the following response, showing you step-by-step what it will be doing and the logic behind it:

First, analyzing the .CSV file
First, analyzing the .CSV file
Then, save the prospects in a document
Then, save the prospects in a document
The preview of the prospects
The preview of the prospects
Done
Done

You can tell the AI agent to tweak it to your specific needs.

Give it specific instructions
Give it specific instructions
The email draft created
The email draft created
Done, do you want to further scale the actions?
Done, do you want to further scale the actions?


Gmail draft result:

Actual draft in Gmail
Actual draft in Gmail


Google document result:

The results automatically saved in Google Docs
The results automatically saved in Google Docs

Give your AI agents quality data & clear instructions, and you're all set

Neat, don't you think? You can now create reports automatically and even create highly personalized emails that are based on accurate data. Gone were the days when you have to manually copy and paste email addresses and read boring Excel spreadsheet to customize your email. Now you can just tell AI agent what you want and it will do it.

Is that all that you can do with it? No, of course. Truly, the limit is your imagination, find your most repetitive and boring task and let the AI agent do it!

Of course, us being at Proxycurl will emphasize the importance of having quality data to feed to your AI agents, so that they can perform their intended tasks to the best that they can.

🚀
At Proxycurl, we have a full suite of API products and a LinkDB database product consisting of close to 500 million profiles that powers many, many amazing applications and use cases, like AI agents.
Sign up for an account and get started now!

Zeha Irawan
Share:

Subscribe to our newsletter

Get the latest news from Proxycurl

Featured Articles

Here’s what we’ve been up to recently.