Caleb Sima, CTO, SPI Dynamics

Caleb Sima It’s just a few days before the RSA Conference 2007, and some vendors refuse
to let themselves get lost in the glut of news from companies that scramble
for everyone’s attention at the show.


Take SPI Dynamics, for example. The Web application security provider
earlier this week unveiled the fruits of a three-year endeavor, dubbed
Phoenix, in which the company whipped up a new architecture.


Forming the architectural backbone of the latest release of the company’s WebInspect 7 scanning software, Phoenix aims to turn the traditional, more passive Web application security model on its ear.


Traditional Web application scanners can’t effectively handle the
new-fangled applications written with AJAX , Flash
, or RSS  that have come to typify Web 2.0.


Moreover, the highly distributed nature of many organizations’ applications
makes it difficult for current Web app scanners to search for
vulnerabilities through multiple servers. It’s like looking for several
needles in several haystacks.


SPI Dynamics officials say WebInspect 7 aims to end the suffering because
Phoenix was designed to work with newer applications to thwart more
sophisticated Web attacks.


Internetnews.com recently caught up with Caleb Sima, co-founder and
CTO of SPI Dynamics. Sima and his development team are responsible for the
conception of Project Phoenix and its incorporation as the intelligent
bedrock for WebInspect 7.


Q: Why did SPI Dynamics decide to create the Phoenix architecture, which
you bill as the first Web application-scanning architecture for Web 2.0


We realized the architecture of the Web scanner today is built for the Web
of 2000. It is obvious to everybody the Web has changed, especially in the
last three years. The Web has just skyrocketed to an entirely different way
of using it. Because of that, the vulnerabilities have gotten to be a little
bit different. They have become much more difficult to find.

In order to
identify these vulnerabilities in Web applications, the way the Web scanner
works today had to be drastically changed. The way that we were doing it was
absolutely insignificant and just wasn’t going to work.


That’s why we started Phoenix. We needed to do the next generation of Web
scanning. We needed to make a product that acts like a human hacker, that
thinks like a human hacker and was flexible enough for us to add these
things in a quick amount of time.


Q: With Phoenix, what are some examples of challenges that WebInspect 7
can address that previous versions of WebInspect couldn’t?


There are a bunch of them. One is time. Because Web applications have gotten
so complex and big, scanning a Web application in the old days, in which you
would crawl even a small Web site, would take an hour or so to complete. And
then you would start auditing to find vulnerabilities.

That’s the way
scanners work today. Now, in order to scan even a small Web site today, with
AJAX and all of the other technologies going on, it takes a lot longer. It’s
much more dynamic. You’re missing things. By the time you finish it, two or
three hours into a crawl, you start the auditing. In the meantime, the
auditor, or the person using the product, sits on his butt and basically
does nothing. What happens is there’s three hours of wasted time, and you have to wait for time for auditing to come up.

But with WebInspect 7, we’ve created a
new simultaneous crawl and audit methodology. As soon as you start crawling
pages, it immediately starts auditing for the types of vulnerabilities
you’re looking for. So, you can get pretty much instantaneous results as
soon as you crawl one page.

That is extremely beneficial to an auditor
because as soon as a vulnerability pops up, they can start working on it.
They can determine whether there are false positives. They can determine
whether it’s something they need to put in a report in a different way. They
can confirm by exploitation, they can go further with the vulnerability.
That allows them to work with the product as its working, saving them a
considerable amount of time.

That doesn’t necessarily sound sexy but it is
amazing how much difference it makes from my perspective as a tester.


Q: Do the Web 2.0 technologies also mean you’re seeing new security
vulnerabilities?


The types of vulnerabilities we find are also going to be different.
Previously with WebInspect, there were Web applications that we could not
crawl. For instance, things with CAPTCHA (Completely
Automated Public Turing test to tell Computers and Humans Apart), things
with multi-factor authentication. In order to do scanning for those, you’d
have to hit something, look at your token and enter in a number and then be
able to somehow get the automated product to do that. It was impossible.

With WebInspect 7, as soon as we see a CAPTCHA prompt, we will pause the
scan, alert the user, he can get out of the CAPTCHA, click OK, and boom, the
product continues. Now what will happen, it’s a very simple and easy method
to do these kinds of things.


The biggest problem of course is JavaScript AJAX. In WebInspect 6, we
basically built a JavaScript engine and were able to plug it in in a
rudimentary form. But this JavaScript engine was really built for Phoenix,
so we are now able to implement it in WebInspect 7 to deal with JavaScript
and AJAX technologies.

WebInspect 7 will go through a Web application like a
user and understand what is going on to its full potential. That’s really
important. Because honestly, if you don’t do that, you’re not going to get
to the Web app and you’re going to miss a bunch of vulnerabilities. A lot of
great features and enhancements get thrown into being able to crawl and
understand a Web application better.


Q: Who do you run into on the competitive landscape as you troll for
customers? Why is WebInspect 7 a more attractive alternative?


Our main competitor is WatchFire, and we have a smaller competitor called
Cenzic. Watchfire copies a lot. From our perspective, we were first with a
Windows product, the first with the technology that we have and the first
with our methodology of auditing Web applications.

Unfortunately, everyone
has copied our technology, basing it on a crawl and audit. We just knew
three years ago that that wasn’t going to hold up. We fundamentally changed
the way the engines worked and the way things got audited. They’re running
off of what I consider old 2000-based architecture. They’re trying to keep
up with the Web 2.0 by adding features and enhancements to an old,
fundamentally flawed architecture. It’s just not going to work. At some
point, they’re going to have to turn around.


Q: What do you expect will be one of the hot topics at RSA this year?


I think browser security will be big. Your biggest interface to the Web is
the browser. People are just starting to catch on to its cross-site scripting
vulnerabilities, which has reached a level of absolutely huge potential.
AJAX has added some real big security concerns.

News Around the Web