Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
Microsoft IT

Debugging Microsoft.com 511

teslatug writes "Channel 9 has an interesting video interview with Chris St.Amand and Jeff Stucky who test and debug Microsoft.com. They reveal some of the big problems they used to face such as recycling processes every 5 minutes due to memory leaks and 32 bit limitations, and being unable to push more than 10 Mbits of data to their datacenters due to Windows' networking stack limitations."
This discussion has been archived. No new comments can be posted.

Debugging Microsoft.com

Comments Filter:
  • 10Mbits/s? really? (Score:2, Interesting)

    by psyon1 ( 572136 ) on Monday December 05, 2005 @09:00PM (#14190146) Homepage
    Is windows really limited to 10mb/s due to the network implementation? Now I am really glad I convince people to use Linux, we have one server pushing 480Mbits/s or so using Lighttpd.
  • by GNUALMAFUERTE ( 697061 ) <almafuerte@@@gmail...com> on Monday December 05, 2005 @09:00PM (#14190147)
    They should be redesigned.

    That's a big problem of software made by companys:

    1 - The company's cashflow is based arround selling new versions of the software
    2 - They can't sell to it's customers improvements that they customers can't see
    3 - There is a fixed time that can go by beetween one release and the next one
    4 - Resources are limited

    Because of this, a major redesign is something that won't be profitable, because only the advanced users will note the changes, but 99% of their customers won't, so the software won't sell well. Bug fixes also won't sell, because they are also unvisible to the naked eye of the majority of the userbase, and also customers expect those changes to be free.
    So, some companys only can expect revenue from a given software once a year, and they have to invest into that software, a given set of limited resources over, say, 6 months, when they have to freeze the featureset so they can start debugging. Seeing which things sell, they will obviously focus their atention on: New Features, and a nicer GUI.
    OTH, a project that doesn't have a company running it, can just get out lots of upgrades, when needed, and focus their time on making the software better, even if some of the changes made to the software won't be seen by most of it's users.

    With software prices dropping, and Free Software proving to be a better option, the budget of software companys will be even more limited, and we won't see this situation changing anytime soon.
  • URLs always change (Score:2, Interesting)

    by hey ( 83763 ) on Monday December 05, 2005 @09:07PM (#14190190) Journal
    I which they could keep their URLs (on microsoft.com) fixed. That way links to the site won't alaways be broken. Its not that hard.
  • by GNUALMAFUERTE ( 697061 ) <almafuerte@@@gmail...com> on Monday December 05, 2005 @09:13PM (#14190221)
    Absolutely true. I used to work for a hosting company, we had GNU/Linux and Windows servers.
    The GNU/Linux servers were the ones with more hits, and the ones that required less atention. The windows servers were a pandora box of problems. IIS just can't hold up by itself, if you just serve static pages you are ok, but when people starts using that asp + odbc shit, you have to restart IIS every 5 fucking minutes. We used to receive a stupid "too many conections" from ODBC in our log, and restarting the stupid services woudln't do a damn thing, all you could do was restart the machine, Yes, restart a SERVER. That's about the worse thing a sysadmin can go through, the panic of not knowing if that crappy windows was going to come back up or not. OTH, our GNU/Linux machines with sites running a variety of CGI apps (PHP, Perl, etc), all using MySQL, supported 5 times the load on the windows machines without complaining, and i'm talking about 300 sites on simple x86 hardware, less powerfull than the one on the windows machines, that died with less than 100 sites ...
  • by DonGar ( 204570 ) on Monday December 05, 2005 @09:27PM (#14190293) Homepage
    The article you point to is really interesting, but it talks about offloading the TCP stack handling to the NIC, not about changes to the wire protocol.

    This is an interesting and powerful technology, however the general concept isn't new. More importantly it's not overcoming limitations in TCP, only limitations in the PC architecture and in OS implementations of TCP.

    MS may be proposing changes to TCP to boost performance, but they don't seem to be covered by the article you are linking. In addition, TCP/IP implementations based on the improvements in various RFC's is perfectly capable of multi-gigabit throughput. I seem to remember reading a slashdot article about Internet 2 researchers seeing sustained long distance transfer rates over FTP (which is usually TCP based) that approached 100 Gb/s.
  • by rbarreira ( 836272 ) on Monday December 05, 2005 @09:29PM (#14190305) Homepage
    According to microsoft, the MSN messenger service (which serves to around 70 million people) used to run on 250 32-bit servers, and now it runs on just 25 or something like that... (apparently one of the big reasons was the limit on the number of tcp connections).

    It's quite amazing to think that a service as huge as messenger can run on just 25 servers!
  • by E-Lad ( 1262 ) on Monday December 05, 2005 @09:41PM (#14190368)
    Wow. Just wow.

    I look at Solaris (err, OpenSolaris) and how it can now push a 10Gb/s interface at line speed [sun.com] (or close to it) and MS has struggled up until recently to get satisfactory speeds above 10Mbit/s ?

    Yet another "how do users/admins accept this as OK" thought going through my head re: Windows internals.
  • Remember Hotmail? (Score:5, Interesting)

    by Anti-Trend ( 857000 ) on Monday December 05, 2005 @09:47PM (#14190400) Homepage Journal
    That wasn't the only time that MS has had to eat humble pie about their server capabilities. Remember the whole Hotmail fiasco in late '97 when Microsoft acquired it? The whole thing was running on UNIX and ran just fine. They tried to replace it with NT servers, and it just couldn't stand up under the weight no matter how much hardware they threw at it. As a result, they had to stick with UNIX for quite a while until they could get Windows to the point where they could even pretend to make any real use of it.

    The following is just hearsay, as I've never actually worked for MS. But a couple of engineer buddies I used to work with did some subcontracting for MS, and they said they deployed a whole lot of internal-facing *nix servers during that period. I tend to believe it, because the MS security guys who taught some seminars I attended wouldn't confirm or deny that they used any Linux internally. If they could have denied it in clean conscience, wouldn't they have done so emphatically?

  • by ThinkFr33ly ( 902481 ) on Monday December 05, 2005 @09:47PM (#14190403)
    IIS just can't hold up by itself, if you just serve static pages you are ok, but when people starts using that asp + odbc shit, you have to restart IIS every 5 fucking minutes.

    That's not because of IIS; it's because of the people writing the ASP apps and stupid admins not configuring IIS correctly. If you have stupid people writing applications, those applications have a tendency of doing stupid things. Combine that with admins who don't properly isolate that applications running on IIS and you've got a recipe for requiring an IIS restart "every 5 fucking minutes".

    Give me 5 minutes and I can write a nice app that takes down Apache no problem. A few infinite loops, perhaps each creating a dozen new database connections and allocating a massive string buffer in memory.

    IIS 6.0 has a lot of features built into it that allow for admins to configure application pools to more effectively isolate applications. You can configure those application pools to recycle automatically given certain criteria (like memory usage, CPU usage, # req/sec, @ req/total, etc.), and the pools are isolated from each other so that if one dies due to a misbehaving application, the other applications on the system are not affected.

    We used to receive a stupid "too many conections" from ODBC in our log, and restarting the stupid services woudln't do a damn thing, all you could do was restart the machine, Yes, restart a SERVER.

    Perhaps that's all you could do, but somebody who spent more than 10 minutes reading about administering IIS would know to recycle the ODBC COM+ application to clear out the connection pool. Then they would find the stupid people writing that crappy applications and fire them, or at least isolate their applications in a separate app pool or worker process. (Dllhost.exe.)

    Spare me the anecdotal stories of your LAMP solutions doing so much better than your Windows solutions. You have absolutely no credibility given your complete ignorance.
  • by GigsVT ( 208848 ) on Monday December 05, 2005 @10:01PM (#14190467) Journal
    You might know this. Why is MacOS so slow at ping? Try it sometime, ping flood a linux box, and you'll probably clear 20,000 pings/sec on a normal system (though some versions of linux are slow too). Ping -f a mac and you might get a couple hundred packets back.

    This is true with OS9 and a lot of OSX boxes it seems.
  • by ThinkFr33ly ( 902481 ) on Monday December 05, 2005 @10:05PM (#14190491)
    Watch the video. That wasn't the problem.

    The problem was connecting two datacenters that were physically seperated by a long distance but connected with a high bandwidth pipe... the TCP protocol has problems with this because of latency issues.

    Read this [microsoft.com] to see how they solved it.
  • by erroneus ( 253617 ) on Monday December 05, 2005 @10:26PM (#14190595) Homepage
    You could scan through all of my old posts for background if you like, but back when NT 4.0 was brand new, I helped to save a failing ISP (for at least the next 6 months or so) by setting up a new mail server to replace the one that was failing ever 2 to 10 minutes. I used a machine with less than half the power and resources of the machine already running... and loaded slackware. I think the kernel was jsut over 1.00 at the time.

    Yeah, "old technology" couldn't do anything better than new stuff like NT right? Come to think of it, there's not a LOT of difference between XP's kernel and NT's from what I understand... a few bug fixes here and there... but basically, it uses the same vulnerable messaging scheme and drivers running at ring-0 and all that. ...I guess I've repeated enough digs on microsoft for one posting...
  • by ergo98 ( 9391 ) on Monday December 05, 2005 @11:19PM (#14190816) Homepage Journal
    Yeah, "old technology" couldn't do anything better than new stuff like NT right? Come to think of it, there's not a LOT of difference between XP's kernel and NT's from what I understand... a few bug fixes here and there... but basically, it uses the same vulnerable messaging scheme and drivers running at ring-0 and all that. ...I guess I've repeated enough digs on microsoft for one posting...

    Drivers generally run in kernel mode in Linux, and most other operating systems for that matter. One of the few that doesn't is QNX.

    In any case, the kernel of Windows has been the slowest moving piece of the platform - because it's a very good kernel. It's a mix of performance of reliability that actually exemplifies a lot of great design techniques (BTW: you should have gone for the gold and mentioned VMS).
  • Re:What the... (Score:3, Interesting)

    by tomhudson ( 43916 ) <barbara,hudson&barbara-hudson,com> on Monday December 05, 2005 @11:31PM (#14190865) Journal

    What I mean is, nothing aside from his own principles stop him from using a Windows PC to watch the video Nice to know that people of proincple don't use Windows :-)

    BTW - this is the SAME company that wants everyone to standardize on their bullshit MS-XML, and lies about how they "won't" sue anyone, when they've already laid the groundwork to do exactly that with 6 loopholes in their bs covenant.

    How do you tell if a Microsoftie is lieing?
    Balmer's lips are moving.

  • by batkiwi ( 137781 ) on Monday December 05, 2005 @11:47PM (#14190938)
    What does a TCP/IP improvement to help high-connection-count low-latency scenario have to do with MS trying to solve the problem of a VERY high latency, single connection scenario?

    Or are you just replying to the 3 sentance summary without any information or knowing what you're talking about?

    Microsoft's problems were with the TCP specifications, which they adhered to TOO closely. From the paper, NOT specific to windows (specific to any fully compliant TCP implementation), "under a 10GBPS link with 100ms delay, it will take roughly one hour for a standard TCP flow to fully utilize the link capacity".

    This is due to how TCP is written to share the connection and not swamp it. MS provided a "workaround" for doing single, highspeed data transfers with high latency.
  • by uodeltasig ( 759920 ) on Tuesday December 06, 2005 @01:18AM (#14191303) Homepage
    I like how 10:14-10:18 zooms in right on Chris's keyboard as he types his password. Just using Windows Media Player on cntl+shift+s takes a lot of the guess work out of the password.
    Especially with a little help [slashdot.org] from our friends from UC-Berkley.

    Also, I like 12:32, "so we'll avoid showing ip address... haha we'll have to cut that part out..." like the large 207.46.16.30 address looking at us in the face and then seconds later the 3 ip addresses in clear view on the right.

    "So we have terminal server access to all the servers in the data center, right.". Right, well I wonder how may of those servers, whose IP addresses we just saw, are attached to Chris's login and password?

    Ready, aim, proxy.
  • Re:What the... (Score:3, Interesting)

    by evilviper ( 135110 ) on Tuesday December 06, 2005 @02:22AM (#14191515) Journal
    Yeah, like Bochs is capable of running mplayer at any decent speed.

    Doesn't have to. You can always convert it to another format, even if you can only do it in a fraction of realtime.

    Do people say you can't watch a video, just because it's downloading to slowly to be streamed in realtime?
  • Re:Easy. (Score:3, Interesting)

    by Octorian ( 14086 ) on Tuesday December 06, 2005 @09:15AM (#14192715) Homepage
    Not only is that a dead link, but running Win32 (x86) codecs on a SPARC is fundamentally IMPOSSIBLE.

    The only way to run them would be through an x86 emulator, which would probably be way too slow to result in actual watchable video.

Stellar rays prove fibbing never pays. Embezzlement is another matter.

Working...