Automating security questionnaire intake to help InfoSec analysts get promoted
Triaging requests was manual and time-consuming—while costing credits, morale, and promotions. Asking, "How might we get faster first drafts?" led to designing an AI agent that a publicly traded enterprise software company implemented to automatically handle requests.
Triaging requests was manual and time-consuming—while costing credits, morale, and promotions. Asking, "How might we get faster first drafts?" led to designing an AI agent that a publicly traded enterprise software company implemented to automatically handle requests.
Company
Timeline
Role
Team
Conveyor
7 weeks
7 weeks
Lead Designer
1 PM, 6 engineers (5 software, 1 UX), 1 more designer
Company
Timeline
Role
Team
Conveyor
7 weeks
7 weeks
Lead Designer
1 PM, 6 engineers (5 software, 1 UX), 1 more designer




Irrelevant requests from sellers extend the questionnaire lifecycle by multiple business days
Irrelevant requests from sellers extend the questionnaire lifecycle by multiple business days
InfoSec analysts complain that sellers upload questionnaire requests that shouldn't be handled by the platform—so they end up spending countless hours triaging a queue.
Costing credits
Since credits in Conveyor are consumed every 100 questions, customers want to be able to control which questionnaires get imported. If a questionnaire doesn't have many questions, they may want to handle it manually or reject it altogether.
Missing SLAs
We heard most frustrations from teams who were triaging 100 requests a month. This high volume combined with the manual process contributed to missed SLAs.
Triaging lowers morale
Manually triaging was not only taking the fun out of the job, but it was taking InfoSec analysts away from higher priority initiatives. As one said:
“To get promoted, I need to get out of the queue.”
Unfortunately, triaging takes precedence. This lowered their team's morale significantly.
InfoSec analysts complain that sellers upload questionnaire requests that shouldn't be handled by the platform—so they end up spending countless hours triaging a queue.
Costing credits
Since credits in Conveyor are consumed every 100 questions, customers want to be able to control which questionnaires get imported. If a questionnaire doesn't have many questions, they may want to handle it manually or reject it altogether.
Missing SLAs
We heard most frustrations from teams who were triaging 100 requests a month. This high volume combined with the manual process contributed to missed SLAs.
Triaging lowers morale
Manually triaging was not only taking the fun out of the job, but it was taking InfoSec analysts away from higher priority initiatives. As one said:
“To get promoted, I need to get out of the queue.”
Unfortunately, triaging takes precedence. This lowered their team's morale significantly.




Automating questionnaire intake frees up analysts to get promoted
Analysts use rules to triage
We learned in user interviews that InfoSec teams have criteria for triaging. This often includes things like: deal size, if there's a signed NDA in place, if the content is relevant, and how many questions there are—to name a few.
We used this as a foundation for our automated solution: an AI agent.
Measuring AI agent's success
Customer outcomes:
30%+ reduction in median SLAs after 30 days of use
50% reduction in hours spent triaging after 30 days of use
Performance indicators:
<5% of rejections are incorrectly rejected
<10% of accepted cases are incorrect
95% of questionnaires don’t have edits to the suggested tags
Analysts use rules to triage
We learned in user interviews that InfoSec teams have criteria for triaging. This often includes things like: deal size, if there's a signed NDA in place, if the content is relevant, and how many questions there are—to name a few.
We used this as a foundation for our automated solution: an AI agent.
Measuring AI agent's success
Customer outcomes:
30%+ reduction in median SLAs after 30 days of use
50% reduction in hours spent triaging after 30 days of use
Performance indicators:
<5% of rejections are incorrectly rejected
<10% of accepted cases are incorrect
95% of questionnaires don’t have edits to the suggested tags




Prioritization to launch in 7 weeks
Launching quickly so we could generate demand and build pipeline meant we needed to make tradeoffs. Here's some we made:
Read-only rules
We learned from customers that triage rules don't change often. With this in mind, we prioritized a read-only approach initially to simplify development and user experience. They could update their rules by contacting our Support team.
Salesforce to start
We identified a likely launch partner at an enterprise company that used Salesforce. Knowing this is a common integration, we decided to start with it—with intentions of launching additional common integrations shortly after launch.
Launching quickly so we could generate demand and build pipeline meant we needed to make tradeoffs. Here's some we made:
Read-only rules
We learned from customers that triage rules don't change often. With this in mind, we prioritized a read-only approach initially to simplify development and user experience. They could update their rules by contacting our Support team.
Salesforce to start
We identified a likely launch partner at an enterprise company that used Salesforce. Knowing this is a common integration, we decided to start with it—with intentions of launching additional common integrations shortly after launch.




AI agent can do more than just triage
As part of this work, we incorporated existing functionality into the AI agent's capabilities.
Not only does it triage, but it answers your questionnaires based on your custom tone and verbosity settings and delegates outstanding questions to subject matter experts so questionnaires are completed faster.
As part of this work, we incorporated existing functionality into the AI agent's capabilities.
Not only does it triage, but it answers your questionnaires based on your custom tone and verbosity settings and delegates outstanding questions to subject matter experts so questionnaires are completed faster.




Building customer trust
Try before you buy
One barrier we identified was getting buy-in. To de-risk this and build trust in the feature, we created a test experience (designed by another designer) to allow customers to see their rules in action before having it in production. This also supported proof of concept opportunities who didn't have integrations set up yet.
Getting closer to the vision
Shortly after we launched, we iterated on the tester design to allow customers to reference their rules while inputing values and made the response feel more like a chat.
Try before you buy
One barrier we identified was getting buy-in. To de-risk this and build trust in the feature, we created a test experience (designed by another designer) to allow customers to see their rules in action before having it in production. This also supported proof of concept opportunities who didn't have integrations set up yet.
Getting closer to the vision
Shortly after we launched, we iterated on the tester design to allow customers to reference their rules while inputing values and made the response feel more like a chat.




Evaluating AI agent's success
Unfortunately, I was let go before I could see if we accomplished our goals.
However, we launched our early-access in 7 weeks. We also successfully implemented the AI agent for a publicly traded enterprise software company after getting buy-in. The agent was successfully triaging requests.
View another project
Unfortunately, I was let go before I could see if we accomplished our goals.
However, we launched our early-access in 7 weeks. We also successfully implemented the AI agent for a publicly traded enterprise software company after getting buy-in. The agent was successfully triaging requests.