I chose Vectara to power the search on docs.qumulo.com after investigating a number of other AI search solutions. The tech is now wired directly into the Qumulo Documentation Portal—in addition to other Qumulo websites—searching every guide published there.
The feature that produces summaries in addition to lists is especially useful (let alone the feature that can produce summaries in the language of the original query—we've yet to implement this). Compared to the previous, keyword-based lookup solution we had, Vectara offers a completely new approach to finding information about deploying and configuring Qumulo Core.
Vectara is a very exciting tool that has incredible potential, and I like the fact that, each time I get stuck or run into the issue with GitHub repos such as `vectara-ingest` or `vectara-answer`, there is always a senior engineer to help resolve a filed GitHub issue, provide example code, or adjust the existing code in the repository—so far, everyone has been very responsive.
I like the fact that Vectara doesn't make me buy into an ecosystem such as AWS, Azure, or GCP. Any resources that Vectara uses are for Vectara to manage, which avoids vendor lock-in and also abstracts all the infrastructural issues from the boots-on-the ground deployment and configuration that the administrator performs. Review collected by and hosted on G2.com.
Vectara has a very steep learning curve for someone who isn't a coder by trade and, while I'm moderately proficient with a variety of development frameworks and tools, I'm a documentarian, not a developer, which means that I prefer a turnkey solution: something to deploy, configure, and adjust within no more than 2-3 days in total. This isn't necessarily a serious concern—as Vectara grows and develops their offering, they'll have a much better definition for the main user personas.
The following might seem like a strange concern, but Vectara currently offers far too many examples and scenarios instead of offering solutions to the most *common* use cases. To close this gap for the time being, I recommend creating a foolproof, end-to-end Getting Started Guide that would give the user everything she needs to implement AI search for the most common scenario: a website on a domain (example.com) or subdomain (docs.example.com) that has a sitemap.
Currently, Vectara's models appear to be based mainly on variously configured flavours of OpenAI's ChatGPT-3.5 and ChatGPT-4. It seems that using the latest AI engines slows down Vectara's queries. While I'm certain that the tuning of the prompts and adjustments given to the models will be improved in short order, it might be nice to combine a variety of engines, especially for the summaries. Review collected by and hosted on G2.com.
Validated through a business email account
Organic review. This review was written entirely without invitation or incentive from G2, a seller, or an affiliate.


