2M New Websites a Year Compromised To Serve Malware 72
SkiifGeek writes "Sophos claims that they are detecting 6,000 new sites daily that have been compromised to serve malware to unsuspecting site visitors, with 80% of site owners not aware that they have been compromised — though this figure is probably on the low side. With increasingly vocal arguments being put forward by security experts criticizing the performance and capability of site validation tools (though many of these experts offer their own tools and services for similar capabilities), and rising levels of blended attacks, perhaps it is time you reviewed the security of your site and what might be hiding in infrequently used directories."
How to Check a LAMP Server? (Score:4, Interesting)
Re: (Score:1)
hit it with a hammer (Score:2)
Re: (Score:1)
The only way to be sure is to run it from a CD image and reboot nightly
"Anyone know of a good way to scan a set up? Sophos says that they are detecting thousands of new sites - how are they scanning them?"
They're not, they just need a little boost to the stock price, please buy our PRODUC~1
Re: (Score:2)
Re: (Score:2)
Re: (Score:1)
Re: (Score:2)
Re: (Score:3, Informative)
Re:How to Check a LAMP Server? (Score:5, Informative)
Might not be the best solution but it should be easy to implement. Larger sites can do incremental scans. It would be harder to detect corruption of databases, though, unless you know what to look for or have a concrete way of validating the contents.
=Smidge=
Re: (Score:2, Informative)
You're right that it won't help you detect that somebody has managed to insert a chunk of javascript or PHP in your insecure mySQL/PHP web app, though. Perhaps a combination of Snort, Ntop (if it wasn't shit), a "hardened" PHP binary and config, and log monitoring would alert you in the case of an attack.
The problem is that there's a lot of badly written or out of date software out there t
Re: (Score:1)
I meant "vulnerable", but feel free to insert jokes about neurotic software here:
Re: (Score:1, Informative)
http://radmind.org/ [radmind.org]
Re: (Score:2, Informative)
Congratulations! You have just described Tripwire [sourceforge.net].
Re: (Score:1)
Tripwire is good for systems that are well understood. If you don't understand the changes that can happen on a system then it won't do you much good. But running Tripwire gives the administrator an understanding of system changes over time. So initial unde
Re: (Score:1)
will scan directories recursively and generate all the md5s into
a single file which you could compare against a baseline:
http://md5deep.sourceforge.net/ [sourceforge.net]
Re: (Score:1)
AIDE - Advanced Intrusion Detection Environment
AIDE is a file integrity checker that supports regular expressions. Licensed with GPL.
www.cs.tut.fi/~rammer/aide.html
Re: (Score:2)
Radmind (Score:2, Informative)
Re: (Score:1)
AFICK. (Score:2)
Also, keep an eye on how big your maillog files are - if they suddenly grow by some exponent, you've been turned into a spam server (or a newsletter went out - five seconds of peeking at the live output should tell you which).
Also, you can keep an eye on the http access l
Additionnal malware detection tools (Score:2)
- chkrootkit [chkrootkit.org]
- rkhunter [rootkit.nl]
They are scripts that scan the system for known root kits, weird behaviours and hidden files in unusual places.
They can both be used to scan an offline system (booted from a live-cd and the system mounted under some directory),
and a live online system (they check the system for suspicious behaviour that may reveal a root-kit trying to hide it self - for example the "ps" co
Re: (Score:2)
The server does NOT run database -- only (pure static) apache. Scripts are NOT run on this machine. Certain things are directed off to https, on another machine, with user/password authentication. That, in turn, actually talks
Re: (Score:1)
Re: (Score:1)
validated certified imdemnified .. (Score:2)
Hmm, time to improve the common tools (Score:4, Interesting)
For example, once you set up your web site, "lock it" so if there are any changes to files or directories that shouldn't change, the site will break in a non-harmful way rather than be compromised.
If and when these files need updating, the "unlock" process should be done using a tool independent of the main web-server process, perhaps by using a different web-server process running on a different port or even a process on a different computer that validates the request then passes it on to the main web server.
Re: (Score:1, Funny)
Re: (Score:1)
Everything on the Web server should be mounted read-only, preferrably from a machine behind a firewall. A firewall sits behind that machine and your inside network. The only way to write to the file system should be from behind the firewall. Any temporary files that need to be created for download or parsing or whatever, where read/write is necessary, should only be done from a RAM disk. Reboot the server nightly.
The database server should also sit on the
too complicated for "simple" setups (Score:1)
However, a stack that puts a virtual or other hardened subsystem to hide the non-read-only files and databases behind in an easy-to-use form should be doable.
Re: (Score:1)
Actual different machines with actual different firewalls are good for hosted solutions and IT departments that know what they are doing, but they are too complicated for non-geek do-it-yourself mom-and-pop-businessman/home-user solutions.
Should the non-geek do-it-yourself mom-and-pop-businessman/home-user REALLY be putting a live box out on the public Intartubes with exposed services? Wouldn't they be MUCH better off with a hosted solution, especially given that shared hosting can cost as little as $5 a month?
Re: (Score:2)
For example, once you set up your web site, "lock it" so if there are any changes to files or directories that shouldn't change, the site will break in a non-harmful way rather than be compromised.
If it's not supposed to change at all, just issue chattr +i on it to make it immutable. Then it won't change, even w/ system root permissions. Just remember to unset the flag any time that you do want to change something ("chattr -i").
Re: (Score:2)
chattr is not meant to be any sort of cure-all - far from it. But for files (and even directories) in the chroot jail that don't change very often (if at all), like logo images, certain site-structural files and scripts, etc? It works just fine.
For instance, a typical Wiki or any sort of CMS rig-up has lots of CGI and/or PHP files that you wouldn't expect to see modified until you either apply a patch or customize them yourself. W
Re: (Score:1)
Just check your access logs (Score:1, Informative)
my 2 cents...
Re: (Score:2)
I would just run a perl script that does a regex on the access logs for anything that does not match the files that should be delivered to clients. Put the perl script in a cron job and let it run.
Sounds like LogCheck. I'm using that; it provides a decent organized daily summary of all access to the server.
Also do an MD5 hash on those files regularly and check for any changes to static files.
Know any pre-written scripts/software that handle this? The trouble is, you'd have to secure the hashes as well, which gets tricky, plus it doesn't offer any help for regularly changing content or database contents (where a content hack like secretly added JS would probably be inserted). But still, better security for non-changing files seems like a good idea; the chattr +i suggestion above se
Re: (Score:2)
virtualized rootkits (Score:3, Interesting)
Re: (Score:1)
Completely useless. (Score:4, Insightful)
I would love to put that list in my squid blocking file to protect my users.
Re: (Score:2)
If it was domains names the same problem would apply but after domains cleaned infected files.
I say all of this because I was a victim of stupid block lists when I got a new IP and tried to send email out on it. It was blocked because of the previous owner and getting removed from most lists was non-obvio
Re: (Score:2)
6000 sites? (Score:1, Interesting)
what does this look like from the client? (Score:4, Interesting)
I tend to think that keeping my OS patched keeps me pretty safe, but there's always a delay after a new vulnerability is discovered before the patches come out (the zero day) and what concerns me is that if someone has a very large network of compromised web servers, they can roll out a zero day vulnerability to all of them and do a lot of damage.
As to symptoms, I think spyware used to be the big problem, and infected computers would have popups and such. But now I think that infected machines will be used primarily to send spam. Is that correct?
Re: (Score:2)
What I wanna know is ... (Score:3, Interesting)
Imagine all the useful things we could do for the world if we all had access to this distributed computing power.
Re: (Score:2)
Re: (Score:2)
Well, I think you might be a bit late with that.
But think of the good things that could be done with a free and open implementation.
OTOH, it's been more than 25 years since the first true distributed OS was announced, and the idea hasn't exactly taken the world by storm.
80% (Score:1, Interesting)
with 80% of site owners not aware that they have been compromised
Wait. So 20% of site owners know their site has been comprimised and they haven't done anything about it and are still serving up malware? Sounds to me like someones making up statistics.
Yes... (Score:3, Interesting)
Re: (Score:1)
Re: (Score:1)
Yes, some do - and it's not a knock on Linux, either.
I work at a small webhost. We're 100% Linux, and have somewhere in the low hundreds of thousands of sites on about forty servers. I come across a compromised site about every other day, and those are the ones that are making themselves obvious - malicious javascript, form abuse, SQL injections, etc. Being on Linux servers has nothing to do with how secure the sites are. The users pick their own passwords and manage their own content, and the sites t
Somebody should warn... (Score:2, Funny)
Vendor FUD or Real? (Score:4, Interesting)
Absent more detail, I am calling shenanigans on this statistic, Sophos, and the Register. I am soooo sick of the FUD.
Harumph!
Re: (Score:1)
Time for a white hat virus? (Score:1)
Re: (Score:2)
Re: (Score:2)
Oh, and it would probably kill most/all of the bot nets.
No -- many of the bots nowadays lock down the PC themselves (once they're in...) to keep it "safe" from competing bots. They even actively remove other bots when they can manage it.
As for the idea, though... I think about that as well. Even if just to get onto computers that haven't been compromised by a really effective bot yet (as I mentioned above) would be a big step.
Alas, most of the people talented enough to write such a thing are probably either:
* well-employed enough that they don't want to risk